TY - GEN
T1 - On the Feasibility of Crawling Linked Data Sets for Reusable Defect Corrections.
AU - Sampaio, Sandra
A2 - Knuth, Magnus
A2 - Kontokostas, Dimitris
A2 - Sack, Harald
A2 - Knuth, Magnus
A2 - Kontokostas, Dimitris
A2 - Sack, Harald
PY - 2014/9/2
Y1 - 2014/9/2
N2 - Current linked open data standards have encouraged thepublication of a large number of data sets on the publicWeb. While some data providers put a lot of energy andresources into maintaining high quality data, others do not,meaning that the quality of the data in many LOD sourcesis variable and unpredictable. This makes the constructionof novel applications on top of the data more difficult andexpensive than it otherwise would be. However, these same data standards also open up possibilitiesfor new ways of managing information quality (IQ). Inthis paper, we propose one such approach, the IQ-bot, andpresent the results of our study of its feasibility. An IQ-botis a 3rd party component that crawls the Web of data,looking for changes that have been made to data sets, andinferring from them where a correction to a data defect hasbeen made. These corrections can then potentially be madeavailable for application to other databases showing evidenceof the presence of the same data defect. In this way, the benefits of the curation effort put into a small number of datasets can be propagated throughout the Web of data.
AB - Current linked open data standards have encouraged thepublication of a large number of data sets on the publicWeb. While some data providers put a lot of energy andresources into maintaining high quality data, others do not,meaning that the quality of the data in many LOD sourcesis variable and unpredictable. This makes the constructionof novel applications on top of the data more difficult andexpensive than it otherwise would be. However, these same data standards also open up possibilitiesfor new ways of managing information quality (IQ). Inthis paper, we propose one such approach, the IQ-bot, andpresent the results of our study of its feasibility. An IQ-botis a 3rd party component that crawls the Web of data,looking for changes that have been made to data sets, andinferring from them where a correction to a data defect hasbeen made. These corrections can then potentially be madeavailable for application to other databases showing evidenceof the presence of the same data defect. In this way, the benefits of the curation effort put into a small number of datasets can be propagated throughout the Web of data.
M3 - Conference contribution
BT - Proceedings of the 1st Workshop on Linked Data Quality co-located with 10th International Conference on Semantic Systems, LDQ@SEMANTiCS 20144
PB - RWTH Aachen University
CY - http://ceur-ws.org/Vol-1215/paper-02.pdf
T2 - LDQ@SEMANTiCS
Y2 - 2 September 2014 through 2 September 2014
ER -