Categories

RSS Aggregator

LoCloud is a Best Practice Network of 32 partners, co-funded under the CIP ICT-PSP Programme of the [...]

This past Monday, in my musing, I mentioned Kanu Hawaii–a nonprofit that recruits people to take act [...]

The iPad market is saturated. Tablets are gadgets for a largish, elite niche. So, as a technology, t [...]

The Art & Science of Curation is a project which explores ideas around Curation and the role of [...]

New Byzantine texts were added to the Thesaurus Linguae Graecae on 22 April 2014 0082 APOLLONIUS DYS [...]

(10) metadata entry Contribution: Susanne Uhlirz Name: Susanne Uhlirz URL: link to the original post [...]

Archäologie und Computer 2007. Workshop 12 Wien 2008. PDF-Files auf CD-ROM Preis: zehn Euro ISBN 978 [...]

The following is an excerpt from a Program Update by Christa Williford, with contributions from Amy [...]

Todays list of Open Access (free to read) Archaeology articles:STAC: The Severe Terrain Archaeologic [...]

Personal Digital Archiving 2014. Photo by Bill Lefurgy. Cinda May, a key organizer of the Personal D [...]

Google has released all its old Google Street View pictures, so we can travel back in time…. We’ve g [...]

New Voices In Classical Reception Studies Conference Proceedings Volume 1 Conference Proceedings Vol [...]

At the Inaugural Texas Digital Humanities Consortium Conference (TXDHC) on April 12, Elijah Meeks su [...]

Irmengard MAYER1 / Marina DÖRING-WILLIAMS1/ Georgios TOUBEKIS2 / Michael JANSEN2 / Michael PETZET3 ( [...]

Filippo SUSCA (Dipartimento di Progettazione dell’Architettura, Facoltà di Architettura di Firenze, [...]

Top Subscribed RSS

Top Contributors

Impressions of the ‘Hadoop-driven digital preservation Hackathon’ in Vienna

More than 20 developers visited the ‘Hadoop-driven digital preservation Hackathon’ in Vienna which took place in the baroque room called “Oratorium” of the Austrian National Library from 2nd to 4th of December 2013. It was really exciting to hear people vividly talking about Hadoop, Pig, Hive, HBase followed by silent phases of concentrated coding accompanied by the background noise of mouse clicks and keyboard typing.

There were Hadoop newbies, people from the SCAPE Project with some knowledge about Apache Hadoop related technologies, and, finally, Jimmy Lin who works currently as an associate professor at the University of Maryland and who was employed as research scientist at Twitter before. There is no doubt that his profound knowledge of using Hadoop in an ‘industrial’ big data context was that certain something of this event.

The topic of this Hackathon was large-scale digital preservation in the web archiving and digital books quality assurance domains.  People from the Austrian National Library presented application scenarios and challenges and introduced the sample data which was provided for both areas on a virtual machine together with a pseudo-distributed Hadoop-installation and some other useful tools from the Apache Hadoop ecosystem.

I am sure that Jimmy’s talk about Hadoop was the reason why so many participants became curious about Apache Pig, a powerful tool which was humorously characterised by Jimmy as the tool for lazy pigs aiming for hassle-free MapReduce. Jimmy gave a live demo running some pig scripts on the cluster at his university explaining how pig can be used to find out which links point to each web page in a web archive data sample from the Library of Congress. Asking about his opinion on Pig and Hive as two alternatives for data science to choose from, I found it interesting that he did not seem to have a strong preference for Pig. If an organisation has a lot of experienced SQL experts, he said, Hive is a very good choice. On the other hand, from the perspective of the data scientist, Pig offers a more flexible, procedural approach for manipulating data and to do data analysis.

Towards the end of the first day we started to split into several groups. People gathered ideas in a brainstorming session which at the end led to several groups:

·  Cropping error detection

·  Full-text search on top of warcbase

·  Hadoop-based Identification and Characterisation

·  OCR Quality

·  PIG User Defined Functions to operate on extracted web content

·  PIG User Defined Functions to operate on METS

Many participants made their first steps in Pig scripting during the event, so it is clear that one cannot expect code that is ready to be used in a production environment, but we can see many points to start from when we do the planning of projects with similar requirements.

On the second day, there was another talk by Jimmy about HBase and his project WarcBase which looks like a very promising approach of providing a scalable HBase storage backend with a very responsive user interface that offers basic functionality of what the WayBack machine does for rendering ARC and WARC web archive container files. In my opinion, the upside of his talk was to see HBase as tremendously powerful database on top of Hadoop’s distributed file system (HDFS), Jimmy brimming over with ideas about possible use cases for scalable content delivery using HBase.  The downside was to hear his experiences about how complex the administration of a large HBase cluster can become. First, additionally to the Hadoop administration tasks, it is necessary to keep additional daemons (ZooKeeper, RegionServer) up and running, and he explained how the need for compacting data stored in HFiles, once you believe that the HBase cluster is well balanced, can lead to what the community calls a “compaction storm” that blows up your cluster – luckily this only manifests itself with endless java stack-traces.

One group provided a full text search for WarcBase and they picked up the core ideas from the developer groups and presentations to build a cutting-edge environment where the web archive content was indexed by the Terrier search engine and the index was enriched with metadata from the Apache Tika mime-type and language detection. There were two ways to add metadata to the index. The first option was to run a pre-processing step that uses Pig user defined function to output the metadata of each document. The second option was to use Apache Tika during indexing to detect both the MimeType and language. In my view, this group has won the price of the fanciest set-up, sharing resources and daemons running on their laptops.         

I was impressed how in the largest working group the outcomes were dynamically shared between developers: One implemented a Pig user defined function (UDF) making use of Apache Tika’s language detection API (see section MIME type detection) which the next developer used in a Pig script for mime type and language detection. Also Alan Akbik, SCAPE project member, computer linguist and Hadoop researcher from the University of Berlin, was reusing building blocks from this group to develop Pig Scripts for old German language analysis using dictionaries as a means to determine the quality of noisy OCRed text. As an experienced Pig scripter he produced impressive results and deservedly won the Hackathon’s competition for the best presentation of outcomes.

The last group was experimenting with functionality of classical digital preservation tools for file format identification, like Apache Tika, Droid, and Unix file, and looking into ways to improve the performance on the Hadoop platform. It’s worth highlighting that digital preservation guru Carl Wilson found a way to replace the command line invocation of unix file in FITS by a Java API invocation which proved to be ways more efficient.

Finally, Roman Graf, researcher and software developer from the Austrian Institute of Technology, took images from the Austrian Books Online project in order to develop python scripts which can be used to detect page cropping errors and which were especially designed to run on a Hadoop platform.

On the last day, we had a panel session with people talking about experiences regarding the day-to-day work with Hadoop clusters and the plans that they have for the future of their cluster infrastructure.

I really enjoyed these three days and I was impressed by the knowledge and ideas that people brought to this event.

(41)

Share
metadata entry

Contribution: shsdev

Name: shsdev

URL: link to the original post

Entry: http://www.openplanetsfoundation.org/blogs/2013-12-06-impressions-%E2%80%98hadoop-driven-digital-preservation-hackathon%E2%80%99-vienna

Language: English

Format: text/html