New Tech Post - data /taxonomy/term/269/0 en DataSift: New Frontiers in Search /2010/12/16/datasift-new-frontiers-in-search <div class="facebookshare-box" style="float:right"><fb:share-button href="/2010/12/16/datasift-new-frontiers-in-search" type =" box_count" ></fb:share-button></div><div class="tweetbutton"><a href="" class="twitter-share-button" data-count="vertical" data-via="newtechpost" data-related="tom_murphy:Writer" data-text="New Tech Post - DataSift: New Frontiers in Search" data-url="/2010/12/16/datasift-new-frontiers-in-search" data-lang="en"></a></div><div class="linkedinbutton"><script type="in/share" data-counter="top" data-url="/2010/12/16/datasift-new-frontiers-in-search"></script></div><p><img src="http://socialmedia.net/sites/socialmedia.net/files/DStitle550.png" /><br /> <b><a href="http://uk.linkedin.com/in/sarahblow">Sarah Blow</a> is a computer science graduate of Manchester University in England. Sarah is now working on bringing <a href="http://datasift.net">DataSift</a> through its <a href="http://en.wikipedia.org/wiki/Software_testing">alpha testing</a> phase in preparation for wider use. The filtering technology underpinning DataSift allows for highly refined searches to be able to take place across a selection of social networking platforms.</b></p> <p>Five years ago, in her spare time, Sarah founded <a href="http://girlgeekdinners.com/about-us/founder/">Girl Geek Dinners</a> to alleviate the isolation that many women may feel while working in what is still a male dominated tech industry.</p> <p><img src ="http://socialmedia.net/sites/socialmedia.net/files/Sarah Blow ava.png" align ="left" hspace ="10" /><b>How did DataSift come to be?</b></p> <p>“DataSift is pretty much the back end system that powers <a href="">TweetMeme</a>. We wanted to rebuild the engine that because we knew we could do more with it. Rather than changing TweetMeme, we created this new brand called DataSift.</p> <p><b>Who is it for?<b></b></b></p> <p>“The types of people we are bringing in are big financial services companies, anyone looking to do marketing and marketing analysis, agencies that are looking after brands and so on. Pretty much anything you can think of where you want to find deeper layers of information. </p> <p>“While it is in alpha we are looking mainly for developers but with the mindset of bringing anyone in who is filtering and creating content. We want to find out what their needs are beyond the basics of what we’ve got there at the moment. </p> <p>"We have basically asked, ”What do people want to do with the system?” And now we are looking at ways of packaging it so it can work in the right ways for different groups of users.”</p> <p><b>What do people use it for?</b></p> <p>“We’ve seen people use it for geo-targeting and geo-mapping content in order to find about particular brands and track them. Also, people use it to find out where their users are based. They can also find the most influential users within their particular market.</p> <ul> <li><b>Use example: San Francisco 49ers </b> <p>"For TechCrunch disrupt we demonstrated the capability of DataSift by using publicly available information from a San Francisco 49ers game that was happening that weekend.</p> <p>"There were three rules which we set up 24 hours in advance." </p> <ul> <li><b>Data Collection:</b> "We first created a base rule. That was pulling in information from everyone from at <a href="http://www.49ers.com/stadium/index.html">Candlestick Park</a>. Anyone who mentioned the name San Francisco 49ers and anyone who mentioned any of the players from the 49ers and the opposing team. <p>"There was no geo data in that in that filtered stream."</p></li> <li><b>Geo-location:</b> "From there we built a second rule on top. Taking all the output from the first rule we said, “Right, now we want anything in San Francisco." So, if someone has set their Twitter location to San Francisco we could pick them up including their tweets about that particular subject but only that subject."</li> <li><b>Geo-targeting:</b> "Then we decided, that’s good, but it’s not perfect. What we really want to know is who is in the Park that’s really seeing the cool stuff. Can we manage to get some twitpics, for example, from inside the Park from people who were actually there? <p>“The only way you can verify they are really in the Park is if they are actually geo-located in the Park. So we have an option on DataSift to do a single point and set a radius around it. We found the geo-target for Candlestick Park and set a one kilometre radius around it which pretty much covers that area and anyone just outside the stadium. </p> <p>"But it didn’t come back with much at all. It literally came back with one user and they hadn’t done any photographs. They had just tweeted that they were there.”</p></li> </ul> <p>That only one user was returned according to the parameters that were set up is very interesting in itself. It would have been reasonably valid to have guessed far more returns. It is always worthwhile to remember that what one assumes about a situation and what really happens may be two entirely different things. </p> <p>This is why having better tools to be able to really drill down into the data and to refine and define the results is so vitally important.</p> </li><li><b>Use example: Starbucks </b> <p>“We were based in San Francisco at the time. So we tried a different exercise using DataSift where we basically said, “We want to find anyone in Starbucks who has got a <a href="http://www.peerindex.net/">PeerIndex</a> score of over 40.” Let’s see who the influential people were in San Francisco at that time and find out which Starbucks they are in today. That was a fun one to do. “</p></li> </ul> <p>“You could do something similar with breaking news. If you knew there was a story breaking in a particular location and you are a news organisation and you want to filter down to find who the actual, legitimate sources were that were actually really in that location using DataSift would certainly be one way of doing it.”</p> <p><b>What is the next step?</b></p> <p>“We are aiming to have a drag and drop interface which we haven’t finished yet. Users who don’t necessarily have a strong technical ability and an understanding of the technicalities of it don’t need to. They shouldn’t need to have to have that level of detail to use the system.</p> <p>“The FSDL language that we have got there we only really expect to be used by developers. It is not really aimed at the general user. But while DataSift is in alpha we’ll teach the general users how to use it in case we take a bit longer doing the other side of it.”</p> <p>The work being done by Sarah and the DataSift team is promising to be a cutting edge development in information retrieval. If you want to help with their alpha testing you can still sign up at <a href="http://datasift.net/">DataSift</a>.</p> Technology data tweetmeme Thu, 16 Dec 2010 15:03:33 +0000 Tom Murphy 244 at ScraperWiki: Hacks and Hackers Day comes to Dublin /2010/10/06/scraperwiki-hacks-and-hackers-day-comes-to-dublin <div class="facebookshare-box" style="float:right"><fb:share-button href="/2010/10/06/scraperwiki-hacks-and-hackers-day-comes-to-dublin" type =" box_count" ></fb:share-button></div><div class="tweetbutton"><a href="" class="twitter-share-button" data-count="vertical" data-via="newtechpost" data-related="tom_murphy:Writer" data-text="New Tech Post - ScraperWiki: Hacks and Hackers Day comes to Dublin" data-url="/2010/10/06/scraperwiki-hacks-and-hackers-day-comes-to-dublin" data-lang="en"></a></div><div class="linkedinbutton"><script type="in/share" data-counter="top" data-url="/2010/10/06/scraperwiki-hacks-and-hackers-day-comes-to-dublin"></script></div><p><a href="http://scraperwiki.com"><img src="http://socialmedia.net/sites/socialmedia.net/files/catsquare.png" /></a><br /><br /> <b><a href="http://blog.scraperwiki.com/2010/09/13/new-event-dublin-hacks-and-hackers-hack-day/">Hacks and Hackers Hack Day</a> is taking place in Ireland on the 16th of November during <a href="http://www.innovationdublin.ie/index.php/about">Dublin Innovation Week</a>. The organiser of the day-long event is <a href="http://scraperwiki.com">ScraperWiki</a>. Their aim is to provide the resources that allow anyone with any kind of programming ability to develop, store, and maintain software tools for the purposes of extracting and linking data.</b></p> <p>By providing the means to create accessibility to data ScraperWiki can allow interested parties such as journalists to take advantage of initiatives such as the UK Government’s policy to make its data more available to the public. Since the <a href="http://en.wikipedia.org/wiki/United_Kingdom_Parliamentary_expenses_scandal">UK Expenses Scandal</a>, where certain British parliamentarians were found to have abused their statutory allowances, journalists have become increasingly aware of the wealth of potential stories that lie in databases around the world. However, this data has usually been stored in a random, unstructured and relatively inaccessible manner.</p> <p><a href="http://twitter.com/#!/ainemcguire"><img src ="http://socialmedia.net/sites/socialmedia.net/files/Ainepicture160.png" align ="left" hspace = "10" /></a>According to <a href="http://twitter.com/#!/ainemcguire">Aine Mcguire</a>, in charge of sales and marketing for ScraperWiki, change has only come recently, “In 2003, a gentleman called <a href="http://en.wikipedia.org/wiki/Julian_Todd">Julian Todd</a> contacted the UK Government to find out how various MPs had voted on the war. When he tried to get this information in order to do some analysis on it he was advised by the <a href="http://www.cabinetoffice.gov.uk">Cabinet Office</a> that all this information was published in <a href="http://www.publications.parliament.uk/pa/cm/cmhansrd.htm">Hansards</a> which is the official publishing body of the UK government. But it was difficult [to access.] It was deep down inside a website and he couldn’t do anything with it. </p> <p>“So Julian went and scraped all that information from Hansards and...then fed it into a website in the UK called <a href="http://www.publicwhip.org.uk">The Public Whip</a> which shows you the voting record of all of the MPs in the UK.</p> <p>“But it was very controversial as he risked imprisonment for doing this because of Crown copyright. But they didn’t imprison him and it was Julian Todd who came up with the idea for ScraperWiki.”</p> <p>Active since March, 2010, Aine says Scraperwiki aims to, “build the largest community supported public data store in the world. </p> <p>“You’ve got <a href="http://en.wikipedia.org/wiki/Main_Page">Wikipedia</a> which supports content that’s predominantly for text and <a href="http://www.openstreetmap.org">OpenStreetMap</a> is for maps. What we want to do is create a wiki for data. We’re taking data that is in a very unstructured style and putting it into our structured data store. Where appropriate we’re adding longitude and latitude tags. We’re geo-tagging it which means that data can be mapped.”</p> <p>In line with its aim of being a worldwide data resource project ScraperWiki has had datasets submitted from countries such as the UK, Brazil, Germany, Estonia, Greece and France to name a few. These datasets cover such subjects as the 11,000 deep sea oil wells around the UK, public transport traffic incidents in London, oil rig accidents and so on.</p> <p>“As well as being a datastore it’s a wiki for code.” Aine explains, “At the moment if you want to do some programming you would go out into the web somewhere, you download some tools, you would install them in the server. Scraperwiki allows you to directly program on the browser so in effect we’re given you lots of libraries for you to program with. </p> <p>“You can write a screenscraper that goes that uses any of the libraries we’ve got in our browser technology. You can use Python PHP, or Ruby. So you can go off and scrape without having to install anything on your PC or server.”</p> <p>An added benefit is that because of the inherently collaborative nature of wikis the possibility exists for code to be updated and improved and shared by other programmers. </p> <p>Aine describes what to expect from the Hacks meet Hackers Hack Day, “At the beginning of the day we have a little presentation about what a Hacks and Hackers Hack Day is all about. Then we give a little presentation on ScraperWiki although we don’t prescribe that they use it. Then we let the journalists and developers gravitate together to form teams over datasets of interests. Then they go off and hack all day. At six o’clock we ask the project groups to come back and present for three minutes each their particular visualisation of the data set that they have worked on.”</p> <p>Prizes are then awarded and there is a reception for the participants to attend. At a previous event in <a href="http://blog.scraperwiki.com/2010/08/11/video-liverpool-hacks-and-hackers-hack-day">Liverpool in July</a> eight projects were produced by journalists and programmers working together using open data.</p> <p>For data driven journalism to flourish information even with the maximum reasonable amount of access granted by governments around the world the problem still exists of data being stored in <a href="http://www.information-management.com/infodirect/20070209/1075772-1.html">data silos</a>. Information has to be accessible not only by other people then those who made the original entries but by other machines as well. Structuring information for greater accessibilty is not going to happen all by itself. It will take the sort of co-ordinated and collaborative effort that organisations such as ScraperWiki offer to really make our world a more open and transparent place to live and work in.<br /> </p><p><i><small>At the moment of writing the Hacks and Hackers day taking place in Dublin is fully subscribed but tickets are still available for the <a href="http://blog.scraperwiki.com/2010/09/24/new-event-hacks-and-hackers-hack-day-belfast-hhhbel/">Belfast event</a> on the 13th of November.</small></i></p> <p> It is a free event and Scraperwiki is a not for profit organisation. Please contact Aine through their <a href="http://scraperwiki.com">website</a> if you would be interested in sponsoring a part of the event.</p> Technology data information openness scraperwiki transparency Wed, 06 Oct 2010 06:06:23 +0000 Tom Murphy 200 at Infographics: Communicating The Essence Of A Tidal Wave Of Information /2010/05/22/infographics-communicating-the-essence-of-a-tidal-wave-of-information <div class="facebookshare-box" style="float:right"><fb:share-button href="/2010/05/22/infographics-communicating-the-essence-of-a-tidal-wave-of-information" type =" box_count" ></fb:share-button></div><div class="tweetbutton"><a href="" class="twitter-share-button" data-count="vertical" data-via="newtechpost" data-related="tom_murphy:Writer" data-text="New Tech Post - Infographics: Communicating The Essence Of A Tidal Wave Of Information" data-url="/2010/05/22/infographics-communicating-the-essence-of-a-tidal-wave-of-information" data-lang="en"></a></div><div class="linkedinbutton"><script type="in/share" data-counter="top" data-url="/2010/05/22/infographics-communicating-the-essence-of-a-tidal-wave-of-information"></script></div><p><img src="http://socialmedia.net/sites/socialmedia.net/files/big_wave_medium.jpg" /></p> <p>As databases around the world begin to share and compare their data with ever-greater meaning and relevance through the rapid roll out of <a href="http://socialmedia.net/linked-data-introduction">Linked Data</a> implementations, it is going to become more and more challenging to corral that data and make it into something user-friendly and practical. After all, data isn't worth anything unless it is usable in some form. </p> <p>We do know that there is a tidal wave of data coming right at us just over the horizon. According <a href="http://www.emc.com/about/news/press/2009/20090518-01.htm">to one source</a> we passed 3 zettabytes (21 zeroes) in 2008. So how so we begin to make sense of it?</p> <p>One answer to help solve the need for increased intelligibility lies in the nascent field of <a href="http://en.wikipedia.org/wiki/Information_graphics">information graphics</a> or infographics for short. Up until now it has been a geeky/arty sub-genre of the Internet and regarded as something quite separate from the hard-core, often macho world of 'real' coding. But researchers, artists, statisticians and folks from all sorts of other fields are realising that not everyone wants to plough through all those numbers and data tables, and why should they, when a simple picture can tell the whole story.</p> <p>But infographics has the possibility of being something far more than the mere prettifying of data. Assembling data in this manner to produce an infographic, a chart, or some other means of communicating an idea visually is really content production. The most important rule of content production is tell a story. That is the secret of all the most interesting infographs.</p> <p>The all-time master (so far) has to be <a href="http://en.wikipedia.org/wiki/Information_graphics">Hans Rosling</a>: it's worth taking a break now and watching <a href="http://www.ted.com/talks/lang/eng/hans_rosling_shows_the_best_stats_you_ve_ever_seen.html">his TED presentation</a> where he sets the record straight on widespread notions concerning the 'developing world'. There is even more of his work over at <a href="http://www.gapminder.org/">gapminder.org</a>.</p> <p>In his historical graphs not only can you view a data subject over time, you can also compare it to other neighbouring data subjects. Plus, in the graphs at Gapminder, you can set your own parameters to achieve a very great degree of fine tuning. It is impossible to play with this data without garnering some very interesting insights into how the world has developed over the last hundred and fifty years or so. Up until very recently, for Hans to have been able to communicate this knowledge and information, which he presents in such an understandable and approachable way, would have required hours, days, weeks, months even, to assemble and put together. Then there would be the time spent writing the book or making the film so that he could share the findings and insights with others.</p> <p>Not only do we have the the chance to make data sensible and easy-to-use, we can - through the application of Linked Data and various new applications - do it in relatively short periods of time. </p> <p>The possibility arises not just of a new and important channel of communication but of a new and exciting possibility of new art form. There will be a great need for more practitioners in this field with the creativity and talent to be able to make huge swathes of data intelligible and useful. </p> <p>In the same way that IT departments devolved into separate computer services and web services departments and we now have a further devolution of social media functions as a professional sector in its own right, I can see data representation as becoming an entire skillset/profession as well.</p> <p>Further reading:</p> <ul> <li><a href="http://queue.acm.org/detail.cfm?id=1805128">This article</a> from the Association of Computing Machinery that was helpful in laying out the land and the terms involved.</li> <li><a href="http://vis.stanford.edu/protovis/">Protovis</a> is an open source programme from Stanford University. It is free and probably a good place to start building your own infographics.</li> <li><a href="http://tables.googlelabs.com/public/tour/tour1.html">Google Fusion Tables</a> is another place you can try building something for free. </li><li><a href="http://nathanyau.com">Nathan Yau</a> (<a href="http://twitter.com/yfd">@yfd on Twitter</a>) has been leading the way for some time with, for instance, the use of UN data to create a <a href="http://nathanyau.com/world-progress-report">world progress report</a>. There is a nice <a href="http://fnewsmagazine.com/wp/2009/01/an-interview-with-nathan-yau/">interview with him in the magazine</a> for the School of the Art Institute in Chicago in which he lays out some of his thinking.</li> </ul> Social Media data Infographics Sat, 22 May 2010 23:15:36 +0000 Tom Murphy 96 at