SlideShare a Scribd company logo
www.kit.edu
@Data Quality Tutorial, September 12, 2016
Crowdsourcing Linked Data Quality Assessment
Amrapali Zaveri
Linked Data - over billion facts
What about the quality?
Motivation - Linked Data Quality
Varying quality of Linked Data sources
Source, Extraction, Integration etc.
Some quality issues require certain interpretation
that can be easily performed by humans
Incompleteness
Incorrectness
Semantic Accuracy
Motivation - Linked Data Quality
Solution: Include human verification in the
process of LD quality assessment via
crowdsourcing
Human Intelligent Tasks (HIT)
Labor market
Monetary Reward/Incentive
Time & Cost effective
Large-scale problem solving approach, divided into
smaller tasks, independently solved by a large
group of people.
Research questions
RQ1: Is it possible to detect quality issues in LD data sets
via crowdsourcing mechanisms?
RQ2: What type of crowd is most suitable for each type of
quality issue?
RQ3: Which types of errors are made by lay users and
experts when assessing RDF triples?
Related work
Crowdsourcing
& Linked Data
Web of data
quality
assessment
Our work
ZenCrowd
Entity resolution
CrowdMAP
Ontology alignment
GWAP for LD
Assessing LD
mappings
(Automatic)
Quality
characteristics of
LD data sources
(Semi-automatic)
DBpedia
WIQA, Sieve,
(Manual)
OUR APPROACH
Find-Verify Phases of Crowdsourcing
Contest
LD Experts
Difficult task
Final prize
Find Verify
Microtasks
Workers
Easy task
Micropayments
TripleCheckMate
[Kontoskostas2013] MTurk
(1) Adapted from [Bernstein2010]
http://mturk.com
LD Experts AMT Workers
Type Contest-based Human Intelligent Tasks
(HITs)
Participants Linked Data (LD) experts Labor market
Task Detect and classify
quality issues in
resources
Detect quality issues in
triples
Reward Most no. of resources
evaluated
Per task/triple
Tool TripleCheckMate Amazon Mechanical Turk,
CrowdFlower etc.
Difference between LD experts & Workers
Methodology
Crowdsource using:
• Linked Data experts - Find Phase
• Amazon Mechanical Turk workers - Verify Phase
Crowdsourcing using Linked Data
Experts — Methodology
Phase I: Creation of quality problem taxonomy
Phase II: Launching a contest
Zaveri et. al. Quality assessment methodologies for Linked Open Data. Semantic
Web Journal, 2015.
D = Detectable, F = Fixable, E = Extraction Framework, M = Mappings Wiki
Crowdsourcing using Linked Data
Experts — Quality Problem Taxonomy
D = Detectable, F = Fixable, E = Extraction Framework, M = Mappings Wiki
Crowdsourcing using Linked Data
Experts — Quality Problem Taxonomy
D = Detectable, F = Fixable, E = Extraction Framework, M = Mappings Wiki
http://nl.dbpedia.org:8080/TripleCheckMate-Demo/
Crowdsourcing using Linked Data
Experts — Contest
Crowdsourcing using Linked Data
Experts — Results
Methodology
Crowdsource using:
• Linked Data experts - Find Phase
• Amazon Mechanical Turk workers - Verify Phase
Crowdsourcing using AMT Workers
Selecting LD quality issues to crowdsource
Designing and generating the micro tasks to present the
data to the crowd
1
2
Dataset
{s p o .}
{s p o .}
Correct
Incorrect +
Quality issue
Steps:
1
2
3
Three categories of quality problems occur
pervasively in DBpedia and can be crowdsourced:
Incorrect/Incomplete object
▪Example: dbpedia:Dave_Dobbyn	dbprop:dateOfBirth	“3”.
Incorrect data type or language tags
▪Example: dbpedia:Torishima_Izu_Islands	foaf:name	“ ”@en.
Incorrect link to “external Web pages”
▪Example: dbpedia:John-Two-Hawks	dbpedia-owl:wikiPageExternalLink			
														<http://cedarlakedvd.com/>
Selecting LD quality issues to
crowdsource
1
Presenting the data to the crowd
• Selection of foaf:name or
rdfs:label to extract human-
readable descriptions
• Values extracted automatically
from Wikipedia infoboxes
• Link to the Wikipedia article via
foaf:isPrimaryTopicOf
• Preview of external pages by
implementing HTML iframe
Microtask interfaces: MTurk tasks
Incorrect object
Incorrect data type or language tag
Incorrect outlink
2
EXPERIMENTAL STUDY
Experimental design
• Crowdsourcing approaches:
• Find stage: Contest with LD experts
• Verify stage: Microtasks
• Creation of a gold standard:
• Two of the authors of this paper (MA, AZ) generated the gold
standard for all the triples obtained from the contest
• Each author independently evaluated the triples
• Conflicts were resolved via mutual agreement
• Metric: precision
Overall results
LD Experts Microtask workers
Number of distinct
participants
50 80
Total time
3 weeks (predefined) 4 days
Total triples evaluated
1,512 1,073
Total cost
~ US$ 400 (predefined) ~ US$ 43
Precision results: Incorrect object task
• MTurk workers can be used to reduce the error rates of LD experts for
the Find stage
• 117 DBpedia triples had predicates related to dates with incorrect/
incomplete values:
”2005 Six Nations Championship” Date 12 .
• 52 DBpedia triples had erroneous values from the source:
”English (programming language)” Influenced by ? .
• Experts classified all these triples as incorrect
• Workers compared values against Wikipedia and successfully classified
this triples as “correct”
Triples compared LD Experts MTurk
(majority voting: n=5)
509 0.7151 0.8977
=
Precision results: Incorrect data type task
Numberoftriples
0
38
75
113
150
Data types
Date Millimetre Number Second Year
Experts TP
Experts FP
Crowd TP
Crowd FP
Triples compared LD Experts MTurk
(majority voting: n=5)
341 0.8270 0.4752
Precision results: Incorrect link task
• We analyzed the 189 misclassifications by the experts:
• The misclassifications by the workers correspond to pages
with a language different from English.
11%
39%
50%
Freebase links
Wikipedia images
External links
Triples compared Baseline LD Experts MTurk
(n=5 majority voting)
223 0.2598 0.1525 0.9412
Final discussion
RQ1: Is it possible to detect quality issues in LD data sets via crowdsourcing
mechanisms?
LD experts - incorrect datatype
AMT workers - incorrect/incomplete object value, incorrect
interlink
RQ2: What type of crowd is most suitable for each type of quality issue?
The effort of LD experts must be applied on those tasks demanding
specific-domain skills. AMT workers were exceptionally good at
performing data comparisons
RQ3: Which types of errors are made by lay users and experts?
Lay users do not have the skills to solve domain-specific tasks, while
experts performance is very low on tasks that demand an extra effort
(e.g., checking an external page)
CONCLUSIONS, CHALLENGES &
FUTURE WORK
Conclusions
A crowdsourcing methodology for LD quality assessment:
Find stage: LD experts
Verify stage: AMT workers
Methodology and tool are generic to be applied to other
scenarios
Crowdsourcing approaches are feasible in detecting the
studied quality issues
Challenges
Lack of gold-standard
Crowdsourcing design — how many workers? how
many tasks? reward?
Microtask design
Future Work
Combining semi-automated and crowdsourcing methods
Predicted vs. crowdsourced metadata
Conducting new experiments (other domains)
Entity, dataset, experimental metadata
Fix/Improve Quality using Crowdsourcing
Find-Fix-Verify Phases
References
Triplecheckmate: A tool for crowdsourcing the quality assessment of
linked data. D Kontokostas, A Zaveri, S Auer, J Lehmann, ISWC 2013.
Crowdsourcing linked data quality assessment. M Acosta, A Zaveri, E
Simperl, D Kontokostas, S Auer, J Lehmann, ISWC 2013.
User-driven quality evaluation of DBpedia. A Zaveri, D Kontokostas, MA
Sherif, L Bühmann, M Morsey, S Auer, J Lehmann, ISEMANTiCS 2013.
Quality assessment for linked data: A survey. A Zaveri, A Rula, A
Maurino, R Pietrobon, J Lehmann, S Auer, Semantic Web Journal 2015.
Detecting Linked Data Quality Issues via Crowdsourcing: A DBpedia
Study. M Acosta, A Zaveri, E Simperl, D Kontokostas, F Flöck, J
Lehmann, Semantic Web Journal 2016.
ACRyLIQ: Leveraging DBpedia for Adaptive Crowdsourcing in Linked
Data Quality Assessment. U Hassan, A Zaveri, E Marx, E Curry, J
Lehmann. EKAW 2016.
Thank You
Questions?
amrapali@stanford.edu

@AmrapaliZ

More Related Content

PDF
Amrapali Zaveri Defense
PDF
Linked Data Quality Assessment: A Survey
PPTX
Crowdsourcing Linked Data Quality Assessment
PPTX
DaCENA Personalized Exploration of Knowledge Graphs Within a Context. Seminar...
PPTX
Semantic Data Retrieval: Search, Ranking, and Summarization
PDF
Data Quality and the FAIR principles
PDF
Workshop on Data Quality Management in Wikidata
PDF
Phd thesis final presentation
Amrapali Zaveri Defense
Linked Data Quality Assessment: A Survey
Crowdsourcing Linked Data Quality Assessment
DaCENA Personalized Exploration of Knowledge Graphs Within a Context. Seminar...
Semantic Data Retrieval: Search, Ranking, and Summarization
Data Quality and the FAIR principles
Workshop on Data Quality Management in Wikidata
Phd thesis final presentation

What's hot (20)

PPTX
A Framework for Linked Data Quality based on Data Profiling and RDF Shape Ind...
PDF
ISWC 2014 Tutorial - Instance Matching Benchmarks for Linked Data
PPTX
Haystack keynote 2019: What is Search Relevance? - Max Irwin
PDF
Stakeholder-centred Identification of Data Quality Issues: Knowledge that Can...
PPTX
Access Lab 2020: Context aware unified institutional knowledge services
PPTX
Recruiting Study Participants Online using Amazon's Mechanical Turk
PPTX
Research Data Sharing and Re-Use: Practical Implications for Data Citation Pr...
PDF
AN EXTENDED DATA OBJECT-DRIVEN APPROACH TO DATA QUALITY EVALUATION: CONTEXTUA...
PPTX
Metadata Quality Assurance
PPTX
PDF
ODSC East 2017: Data Science Models For Good
PPTX
Using the Micropublications ontology and the Open Annotation Data Model to re...
PPTX
SocialCite makes its debut at the HighWire Press meeting
PPTX
Data Discovery and Visualization
PDF
PPTX
Metadata Quality Assurance Part II. The implementation begins
PDF
Lec 1 integrating data science and data analytics in various research thrust
PPTX
Prof. Melinda Laituri, Colorado State University | Map Data Integrity | SotM ...
PPTX
Haystack 2019 - Natural Language Search with Knowledge Graphs - Trey Grainger
PDF
JIST2015-Computing the Semantic Similarity of Resources in DBpedia for Recomm...
A Framework for Linked Data Quality based on Data Profiling and RDF Shape Ind...
ISWC 2014 Tutorial - Instance Matching Benchmarks for Linked Data
Haystack keynote 2019: What is Search Relevance? - Max Irwin
Stakeholder-centred Identification of Data Quality Issues: Knowledge that Can...
Access Lab 2020: Context aware unified institutional knowledge services
Recruiting Study Participants Online using Amazon's Mechanical Turk
Research Data Sharing and Re-Use: Practical Implications for Data Citation Pr...
AN EXTENDED DATA OBJECT-DRIVEN APPROACH TO DATA QUALITY EVALUATION: CONTEXTUA...
Metadata Quality Assurance
ODSC East 2017: Data Science Models For Good
Using the Micropublications ontology and the Open Annotation Data Model to re...
SocialCite makes its debut at the HighWire Press meeting
Data Discovery and Visualization
Metadata Quality Assurance Part II. The implementation begins
Lec 1 integrating data science and data analytics in various research thrust
Prof. Melinda Laituri, Colorado State University | Map Data Integrity | SotM ...
Haystack 2019 - Natural Language Search with Knowledge Graphs - Trey Grainger
JIST2015-Computing the Semantic Similarity of Resources in DBpedia for Recomm...
Ad

Viewers also liked (20)

PPTX
Data Science for the Win
PDF
Managing Completeness of Web Data
PPTX
Survey on Common Strategies of Vocabulary Reuse in Linked Open Data Modeling ...
PPT
A few metrics about Open Data in the cultural sector
KEY
Publishing Linked Open Data in 15 minutes
PPT
Asug Gov Sig Data Quality Metrics Report Sapphire 2008
PPTX
Enhancing educational data quality in heterogeneous learning contexts using p...
PPTX
Linked Data Quality Assessment – daQ and Luzzu
PPTX
Leveraging DBpedia for Adaptive Crowdsourcing in Linked Data Quality Assessment
PDF
Introduction to open data quality et
PDF
An introduction to Linked (Open) Data
PPTX
Open data quality
PDF
Overview of Open Data, Linked Data and Web Science
PDF
Linked Open Data Principles, Technologies and Examples
PDF
Evaluating the possibilities of DataCite for developing 'Open data metrics' o...
PDF
Llinked open data training for EU institutions
PPTX
Phil Ritchie | Putting Standards into Action: Multilingual and Semantic Enric...
PPTX
OWL-based validation by Gavin Mendel Gleasonand Bojan Bozic, Trinity College,...
PPTX
Bio2RDF : A biological knowledge base for the Semantic Web
PPTX
Generating Biomedical Hypotheses Using Semantic Web Technologies
Data Science for the Win
Managing Completeness of Web Data
Survey on Common Strategies of Vocabulary Reuse in Linked Open Data Modeling ...
A few metrics about Open Data in the cultural sector
Publishing Linked Open Data in 15 minutes
Asug Gov Sig Data Quality Metrics Report Sapphire 2008
Enhancing educational data quality in heterogeneous learning contexts using p...
Linked Data Quality Assessment – daQ and Luzzu
Leveraging DBpedia for Adaptive Crowdsourcing in Linked Data Quality Assessment
Introduction to open data quality et
An introduction to Linked (Open) Data
Open data quality
Overview of Open Data, Linked Data and Web Science
Linked Open Data Principles, Technologies and Examples
Evaluating the possibilities of DataCite for developing 'Open data metrics' o...
Llinked open data training for EU institutions
Phil Ritchie | Putting Standards into Action: Multilingual and Semantic Enric...
OWL-based validation by Gavin Mendel Gleasonand Bojan Bozic, Trinity College,...
Bio2RDF : A biological knowledge base for the Semantic Web
Generating Biomedical Hypotheses Using Semantic Web Technologies
Ad

Similar to Crowdsourcing Linked Data Quality Assessment (20)

PDF
Crowdsourcing the Semantic Web
PDF
RDF Data Quality Assessment - connecting the pieces
PDF
Crowdsourced query augmentation through the semantic discovery of domain spec...
PDF
customized eager lazy data cleansing for satisfactory big data veracity
PDF
How to clean data less through Linked (Open Data) approach?
PPTX
Data preparation and processing chapter 2
PDF
Barga Data Science lecture 2
PPTX
Advances in Exploratory Data Analysis, Visualisation and Quality for Data Cen...
PDF
Data Quality
PPT
Thesis Presentation
PDF
My experiment
PPTX
ALIGNED Data Curation Methods and Tools
PPTX
Data Science Introduction: Concepts, lifecycle, applications.pptx
PDF
Evaluation Challenges in Using Generative AI for Science & Technical Content
PDF
Best Practices in Recommender System Challenges
PPT
LinkedUp - Linked Data & Education
PDF
The LDBC Social Network Benchmark Interactive Workload - SIGMOD 2015
PDF
Machinr Learning and artificial_Lect1.pdf
PPTX
PhD Consortium ADBIS presetation.
PPTX
Metadata quality Assurance Framework at QQML2016 - short
Crowdsourcing the Semantic Web
RDF Data Quality Assessment - connecting the pieces
Crowdsourced query augmentation through the semantic discovery of domain spec...
customized eager lazy data cleansing for satisfactory big data veracity
How to clean data less through Linked (Open Data) approach?
Data preparation and processing chapter 2
Barga Data Science lecture 2
Advances in Exploratory Data Analysis, Visualisation and Quality for Data Cen...
Data Quality
Thesis Presentation
My experiment
ALIGNED Data Curation Methods and Tools
Data Science Introduction: Concepts, lifecycle, applications.pptx
Evaluation Challenges in Using Generative AI for Science & Technical Content
Best Practices in Recommender System Challenges
LinkedUp - Linked Data & Education
The LDBC Social Network Benchmark Interactive Workload - SIGMOD 2015
Machinr Learning and artificial_Lect1.pdf
PhD Consortium ADBIS presetation.
Metadata quality Assurance Framework at QQML2016 - short

More from Amrapali Zaveri, PhD (12)

PDF
ESOF Panel 2018
PDF
CrowdED: Guideline for optimal Crowdsourcing Experimental Design
PDF
MetaCrowd: Crowdsourcing Gene Expression Metadata Quality Assessment
PDF
smartAPI: Towards a more intelligent network of Web APIs
PDF
Introduction to Bio SPARQL
PDF
LDQ 2014 DQ Methodology
PDF
TripleCheckMate
PDF
Towards Biomedical Data Integration for Analyzing the Evolution of Cognition
PDF
User-driven Quality Evaluation of DBpedia
PDF
Converting GHO to RDF
KEY
ReDD-Observatory
ESOF Panel 2018
CrowdED: Guideline for optimal Crowdsourcing Experimental Design
MetaCrowd: Crowdsourcing Gene Expression Metadata Quality Assessment
smartAPI: Towards a more intelligent network of Web APIs
Introduction to Bio SPARQL
LDQ 2014 DQ Methodology
TripleCheckMate
Towards Biomedical Data Integration for Analyzing the Evolution of Cognition
User-driven Quality Evaluation of DBpedia
Converting GHO to RDF
ReDD-Observatory

Recently uploaded (20)

PPTX
20250228 LYD VKU AI Blended-Learning.pptx
PDF
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
PDF
Architecting across the Boundaries of two Complex Domains - Healthcare & Tech...
PDF
Empathic Computing: Creating Shared Understanding
PDF
Diabetes mellitus diagnosis method based random forest with bat algorithm
PDF
Peak of Data & AI Encore- AI for Metadata and Smarter Workflows
PDF
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
PPT
Teaching material agriculture food technology
PPTX
Understanding_Digital_Forensics_Presentation.pptx
PPTX
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
PDF
MIND Revenue Release Quarter 2 2025 Press Release
PDF
Network Security Unit 5.pdf for BCA BBA.
PPTX
sap open course for s4hana steps from ECC to s4
PDF
Electronic commerce courselecture one. Pdf
PDF
Encapsulation_ Review paper, used for researhc scholars
DOCX
The AUB Centre for AI in Media Proposal.docx
PPTX
Cloud computing and distributed systems.
PDF
Approach and Philosophy of On baking technology
PDF
The Rise and Fall of 3GPP – Time for a Sabbatical?
PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
20250228 LYD VKU AI Blended-Learning.pptx
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
Architecting across the Boundaries of two Complex Domains - Healthcare & Tech...
Empathic Computing: Creating Shared Understanding
Diabetes mellitus diagnosis method based random forest with bat algorithm
Peak of Data & AI Encore- AI for Metadata and Smarter Workflows
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
Teaching material agriculture food technology
Understanding_Digital_Forensics_Presentation.pptx
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
MIND Revenue Release Quarter 2 2025 Press Release
Network Security Unit 5.pdf for BCA BBA.
sap open course for s4hana steps from ECC to s4
Electronic commerce courselecture one. Pdf
Encapsulation_ Review paper, used for researhc scholars
The AUB Centre for AI in Media Proposal.docx
Cloud computing and distributed systems.
Approach and Philosophy of On baking technology
The Rise and Fall of 3GPP – Time for a Sabbatical?
Agricultural_Statistics_at_a_Glance_2022_0.pdf

Crowdsourcing Linked Data Quality Assessment

  • 1. www.kit.edu @Data Quality Tutorial, September 12, 2016 Crowdsourcing Linked Data Quality Assessment Amrapali Zaveri
  • 2. Linked Data - over billion facts What about the quality?
  • 3. Motivation - Linked Data Quality Varying quality of Linked Data sources Source, Extraction, Integration etc. Some quality issues require certain interpretation that can be easily performed by humans Incompleteness Incorrectness Semantic Accuracy
  • 4. Motivation - Linked Data Quality Solution: Include human verification in the process of LD quality assessment via crowdsourcing Human Intelligent Tasks (HIT) Labor market Monetary Reward/Incentive Time & Cost effective Large-scale problem solving approach, divided into smaller tasks, independently solved by a large group of people.
  • 5. Research questions RQ1: Is it possible to detect quality issues in LD data sets via crowdsourcing mechanisms? RQ2: What type of crowd is most suitable for each type of quality issue? RQ3: Which types of errors are made by lay users and experts when assessing RDF triples?
  • 6. Related work Crowdsourcing & Linked Data Web of data quality assessment Our work ZenCrowd Entity resolution CrowdMAP Ontology alignment GWAP for LD Assessing LD mappings (Automatic) Quality characteristics of LD data sources (Semi-automatic) DBpedia WIQA, Sieve, (Manual)
  • 8. Find-Verify Phases of Crowdsourcing Contest LD Experts Difficult task Final prize Find Verify Microtasks Workers Easy task Micropayments TripleCheckMate [Kontoskostas2013] MTurk (1) Adapted from [Bernstein2010] http://mturk.com
  • 9. LD Experts AMT Workers Type Contest-based Human Intelligent Tasks (HITs) Participants Linked Data (LD) experts Labor market Task Detect and classify quality issues in resources Detect quality issues in triples Reward Most no. of resources evaluated Per task/triple Tool TripleCheckMate Amazon Mechanical Turk, CrowdFlower etc. Difference between LD experts & Workers
  • 10. Methodology Crowdsource using: • Linked Data experts - Find Phase • Amazon Mechanical Turk workers - Verify Phase
  • 11. Crowdsourcing using Linked Data Experts — Methodology Phase I: Creation of quality problem taxonomy Phase II: Launching a contest
  • 12. Zaveri et. al. Quality assessment methodologies for Linked Open Data. Semantic Web Journal, 2015. D = Detectable, F = Fixable, E = Extraction Framework, M = Mappings Wiki Crowdsourcing using Linked Data Experts — Quality Problem Taxonomy
  • 13. D = Detectable, F = Fixable, E = Extraction Framework, M = Mappings Wiki Crowdsourcing using Linked Data Experts — Quality Problem Taxonomy D = Detectable, F = Fixable, E = Extraction Framework, M = Mappings Wiki
  • 15. Crowdsourcing using Linked Data Experts — Results
  • 16. Methodology Crowdsource using: • Linked Data experts - Find Phase • Amazon Mechanical Turk workers - Verify Phase
  • 17. Crowdsourcing using AMT Workers Selecting LD quality issues to crowdsource Designing and generating the micro tasks to present the data to the crowd 1 2 Dataset {s p o .} {s p o .} Correct Incorrect + Quality issue Steps: 1 2 3
  • 18. Three categories of quality problems occur pervasively in DBpedia and can be crowdsourced: Incorrect/Incomplete object ▪Example: dbpedia:Dave_Dobbyn dbprop:dateOfBirth “3”. Incorrect data type or language tags ▪Example: dbpedia:Torishima_Izu_Islands foaf:name “ ”@en. Incorrect link to “external Web pages” ▪Example: dbpedia:John-Two-Hawks dbpedia-owl:wikiPageExternalLink <http://cedarlakedvd.com/> Selecting LD quality issues to crowdsource 1
  • 19. Presenting the data to the crowd • Selection of foaf:name or rdfs:label to extract human- readable descriptions • Values extracted automatically from Wikipedia infoboxes • Link to the Wikipedia article via foaf:isPrimaryTopicOf • Preview of external pages by implementing HTML iframe Microtask interfaces: MTurk tasks Incorrect object Incorrect data type or language tag Incorrect outlink 2
  • 21. Experimental design • Crowdsourcing approaches: • Find stage: Contest with LD experts • Verify stage: Microtasks • Creation of a gold standard: • Two of the authors of this paper (MA, AZ) generated the gold standard for all the triples obtained from the contest • Each author independently evaluated the triples • Conflicts were resolved via mutual agreement • Metric: precision
  • 22. Overall results LD Experts Microtask workers Number of distinct participants 50 80 Total time 3 weeks (predefined) 4 days Total triples evaluated 1,512 1,073 Total cost ~ US$ 400 (predefined) ~ US$ 43
  • 23. Precision results: Incorrect object task • MTurk workers can be used to reduce the error rates of LD experts for the Find stage • 117 DBpedia triples had predicates related to dates with incorrect/ incomplete values: ”2005 Six Nations Championship” Date 12 . • 52 DBpedia triples had erroneous values from the source: ”English (programming language)” Influenced by ? . • Experts classified all these triples as incorrect • Workers compared values against Wikipedia and successfully classified this triples as “correct” Triples compared LD Experts MTurk (majority voting: n=5) 509 0.7151 0.8977
  • 24. = Precision results: Incorrect data type task Numberoftriples 0 38 75 113 150 Data types Date Millimetre Number Second Year Experts TP Experts FP Crowd TP Crowd FP Triples compared LD Experts MTurk (majority voting: n=5) 341 0.8270 0.4752
  • 25. Precision results: Incorrect link task • We analyzed the 189 misclassifications by the experts: • The misclassifications by the workers correspond to pages with a language different from English. 11% 39% 50% Freebase links Wikipedia images External links Triples compared Baseline LD Experts MTurk (n=5 majority voting) 223 0.2598 0.1525 0.9412
  • 26. Final discussion RQ1: Is it possible to detect quality issues in LD data sets via crowdsourcing mechanisms? LD experts - incorrect datatype AMT workers - incorrect/incomplete object value, incorrect interlink RQ2: What type of crowd is most suitable for each type of quality issue? The effort of LD experts must be applied on those tasks demanding specific-domain skills. AMT workers were exceptionally good at performing data comparisons RQ3: Which types of errors are made by lay users and experts? Lay users do not have the skills to solve domain-specific tasks, while experts performance is very low on tasks that demand an extra effort (e.g., checking an external page)
  • 28. Conclusions A crowdsourcing methodology for LD quality assessment: Find stage: LD experts Verify stage: AMT workers Methodology and tool are generic to be applied to other scenarios Crowdsourcing approaches are feasible in detecting the studied quality issues
  • 29. Challenges Lack of gold-standard Crowdsourcing design — how many workers? how many tasks? reward? Microtask design
  • 30. Future Work Combining semi-automated and crowdsourcing methods Predicted vs. crowdsourced metadata Conducting new experiments (other domains) Entity, dataset, experimental metadata Fix/Improve Quality using Crowdsourcing Find-Fix-Verify Phases
  • 31. References Triplecheckmate: A tool for crowdsourcing the quality assessment of linked data. D Kontokostas, A Zaveri, S Auer, J Lehmann, ISWC 2013. Crowdsourcing linked data quality assessment. M Acosta, A Zaveri, E Simperl, D Kontokostas, S Auer, J Lehmann, ISWC 2013. User-driven quality evaluation of DBpedia. A Zaveri, D Kontokostas, MA Sherif, L Bühmann, M Morsey, S Auer, J Lehmann, ISEMANTiCS 2013. Quality assessment for linked data: A survey. A Zaveri, A Rula, A Maurino, R Pietrobon, J Lehmann, S Auer, Semantic Web Journal 2015. Detecting Linked Data Quality Issues via Crowdsourcing: A DBpedia Study. M Acosta, A Zaveri, E Simperl, D Kontokostas, F Flöck, J Lehmann, Semantic Web Journal 2016. ACRyLIQ: Leveraging DBpedia for Adaptive Crowdsourcing in Linked Data Quality Assessment. U Hassan, A Zaveri, E Marx, E Curry, J Lehmann. EKAW 2016.