On the Feasibility of Crawling Linked Data Sets for Reusable Defect Corrections.

Sandra Sampaio, Magnus Knuth (Editor), Dimitris Kontokostas (Editor), Harald Sack (Editor)

    Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

    43 Downloads (Pure)

    Abstract

    Current linked open data standards have encouraged thepublication of a large number of data sets on the publicWeb. While some data providers put a lot of energy andresources into maintaining high quality data, others do not,meaning that the quality of the data in many LOD sourcesis variable and unpredictable. This makes the constructionof novel applications on top of the data more difficult andexpensive than it otherwise would be. However, these same data standards also open up possibilitiesfor new ways of managing information quality (IQ). Inthis paper, we propose one such approach, the IQ-bot, andpresent the results of our study of its feasibility. An IQ-botis a 3rd party component that crawls the Web of data,looking for changes that have been made to data sets, andinferring from them where a correction to a data defect hasbeen made. These corrections can then potentially be madeavailable for application to other databases showing evidenceof the presence of the same data defect. In this way, the benefits of the curation effort put into a small number of datasets can be propagated throughout the Web of data.
    Original languageEnglish
    Title of host publicationProceedings of the 1st Workshop on Linked Data Quality co-located with 10th International Conference on Semantic Systems, LDQ@SEMANTiCS 20144
    EditorsMagnus Knuth, Dimitris Kontokostas, Harald Sack
    Place of Publicationhttp://ceur-ws.org/Vol-1215/paper-02.pdf
    PublisherRWTH Aachen University
    Publication statusPublished - 2 Sept 2014
    EventLDQ@SEMANTiCS - Leipzig, Germany
    Duration: 2 Sept 20142 Sept 2014

    Conference

    ConferenceLDQ@SEMANTiCS
    CityLeipzig, Germany
    Period2/09/142/09/14

    Fingerprint

    Dive into the research topics of 'On the Feasibility of Crawling Linked Data Sets for Reusable Defect Corrections.'. Together they form a unique fingerprint.

    Cite this