A Look Inside the Think Tank...Work by Thomas Steiner
Last week, I attended the 13th International Semantic Web Conference in Riva del Garda, Italy. Google was a gold sponsor, and Vice President Prabhakar Raghavan delivered one of the keynotes. This is my trip report with personal highlights and key take-aways.
I started the conference on Sunday with the Developers Workshop, where I had two papers. The workshop was the first of its kind and was put together by my good friend Ruben Verborgh. It pulled more than 70 people in the room and the workshop was prominently featured during the main conference's opening ceremony.Dandelion. Liepi? et al. showed an ontology visualizer called OWLGrEd. Ebner et al. showed a system called LDcache that deals with caching flaky Linked Data sources. With XSPARQL, DellAglio et al. presented a language and implementation combining XML, SPARQL, and SQL to query heterogeneous data sources. Matteis et al. showed how App Engine or Google Code among others can be used as "free" and queryable triple pattern data stores. Ceccarelli showed an entity linking framework called Dexter. My first contribution is titled Comprehensive Wikipedia Monitoring for Global and Realtime Natural Disaster Detection and focuses on natural disaster detection and monitoring with Wikipedia and online social networks. My second contribution is a paper called Self-Contained Semantic Hypervideos Using Web Components and introduces Web Components for the creation of hypervideos. Consuming Linked Data workshop. The most interesting paper for me was by Rula et al., which dealt with the recency of facts in DBpedia. In the afternoon, I switched to the NLP and DBpedia workshop where the highlight was an amazing 300 slides in 30 minutes keynote by Roberto Navigli on BabelNet, Babelfy, Games with a Purpose, and the Wikipedia Bitaxonomy. Further of interest was a paper by Weisenburger et al. on mining historical data for DBpedia via Wikipedia infoboxes.
Tuesday started with Prabhakar's well-received keynote, in which he provided an overview of search engine development in the last years. His book has a nice summary. I then went to the NLP & IEs track, where the best-paper-award-winning paper on the AGDISTIS framework on entity disambiguation by Usbeck et al. was presented. In the afternoon, I attended the Data Integration and Link Discovery track. I liked a paper by Erxleben et al. that described the integration of Wikidata in the Linked Data Web. From the demos in the evening, I want to specially highlight the best-demo-award-winning paper by my friends Verborgh et al. on Linked Data Fragments on a Raspberry Pie. In general, Linked Data Fragments were one of the themes at this conference with several works citing them and also the release of the official DBpedia Linked Data Fragments interface.paper by Uchida et al. who presented a Chrome extension on browser personalization. I further liked a paper by Khamkham et al. on the CrowdTruth framework for harnessing disagreement in gathering annotated data. In the afternoon, my personal highlight was Verborgh et al.'s full paper on Linked Data Fragments.
I skipped Thursday morning and was back in the afternoon for the Linked Data track. Notable papers include Beek et al.'s LOD Laundromat that provides a solution for streamlining access to Linked Data sources by cleansing and format conversion and Patel-Schneider's analysis of Schema.org and some (author's view) recommendations on how to improve it. Meusel et al. gave an overview of the current state of WebDataCommons project that examines Microdata, RDFa, and Microformats distribution in the CommonCrawl corpus.
PhD thesis successfully defended
I have finally defended my PhD thesis. A raw, unedited recording of the defense is available on YouTube
You can check out my slide deck that I used on http://tomayac.com/phd and the PDF of the thesis itself is available at http://tomayac.com/phd/thesis.pdf. The source code of the thesis is available in the GitHub repository https://github.com/tomayac/phd. I guess this makes me officially Dr. Thomas Steiner from now on.
Weaving the Web(VTT) of Data
This week, I'm attending the World Wide Web conference (WWW2014) in Seoul, Korea. Yesterday, I co-ran the 5th International Workshop on Web APIs and RESTful Design (WS-REST2014). It was a great workshop, for the papers, the keynote by my fellow Google colleague Sam Goto, and above all, for the people:
Now today, I presented work of ours in the workshop Linked Data on the Web (LDOW2014). The title of our paper is Weaving the Web(VTT) of Data, you can see the slides that I used for my talk below.
Telling Breaking News Stories from Wikipedia with Social Multimedia: A Case Study of the 2014 Winter OlympicsCreated on Wednesday, April 02, 2014 at 08:57:09 and categorized as Technical by Thomas Steiner
Telling Breaking News Stories from Wikipedia with Social Multimedia: A Case Study of the 2014 Winter Olympics
This week, I am attending the International Conference on Multimedia Retrieval (ICMR 2014) in Glasgow, Scotland. I contributed a paper to the workshop on Social Multimedia and Storytelling (SoMuS 2014) based on my work around combining Wikipedia Live Monitor and Social Media Illustrator. The title of the paper is Telling Breaking News Stories from Wikipedia with Social Multimedia: A Case Study of the 2014 Winter Olympics, you can read it on arXiv.org (deeplink to the PDF). My slides are available online and also embedded below.
This research was also featured by the MIT Technology Review in an article titled The Evolution of Automated Breaking News Stories.
Bots vs. Wikipedians?Who edits more?
I have just released an app called Bots vs. Wikipedians that displays live stats on Wikipedia and Wikidata bot activity. Bots vs. Wikipedians has a public API that sends out Wikipedia and Wikidata edits as Server-Sent Events. You can learn everything about Server-Sent Events (SSE) from an amazing HTML5 Rocks article by Eric Bidelman. The code of Bots vs. Wikipedians is available under the terms of the Apache 2.0 license and published on GitHub.