SlideShare a Scribd company logo
Link Discovery Tutorial
Benchmarking for Instance Matching Systems
Axel-Cyrille Ngonga Ngomo(1)
, Irini Fundulaki(2)
, Mohamed Ahmed Sherif(1)
(1) Institute for Applied Informatics, Germany
(2) FORTH, Greece
October 18th, 2016
Kobe, Japan
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 1 / 36
The Questions(s)
Instance matching research has led to the development of
various systems.
What are the problems that I wish
to solve?
What are the relevant key
performance indicators?
What is the behavior of the existing
engines w.r.t. the key performance
indicators?
Which are the tool(s) that I should use for my data
and for my use case?
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 2 / 36
Importance of Benchmarking
Benchmarks exist
To allow adequate measurements of systems
To provide evaluation of engines for real (or close to real) use cases
Provide help
Designers and Developers to assess the performance of their tools
Users to compare the different available tools and evaluate suitability for their
needs
Researchers to compare their work to others
Leads to improvements:
Vendors can improve their technology
Researchers can address new challenges
Current benchmark design can be improved to cover new necessities and
application domains
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 3 / 36
The Answer: Benchmark your engines!
Instance Matching/Linking Benchmark comprises of
Datasets: The raw material of the benchmarks. These are the source and
the target dataset that will be matched together to find the links between
resources
Test Cases: Address heterogeneities (structural, value, semantic) of the
datasets to be matched
Gold Standard (Ground Truth / Reference Alignment): The "correct
answer sheet" used to judge the completeness and soundness of the instance
matching algorithms
Metrics: The performance metric(s) that determine the systems behaviour
and performance
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 4 / 36
Benchmark Datasets: Characteristics
Nature
Real Datasets: Widely used datasets from a domain of interest
+ Realistic conditions for heterogeneity problems
+ Realistic distributions
- Error prone, hard to create Reference Alignment
Synthetic Datasets: Produced with a data generator (that hopefully produces
data with interesting characteristics
+ Fully controlled test conditions
+ Accurate, Easy to create Reference Alignments
- Unrealistic distributions
- Systematic heterogeneity problems
Schema
Datasets to be matched have the same or different schemas
Domain
Datasets come from the same or different domains
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 5 / 36
Benchmark Test Cases: Variations
Value
Name style abbreviations, Typographical errors, change format
(date/gender/number), synonym change, language change (multilinguality)
Structural
Change property depth, Delete/add property, split property values,
transformation of object/data to data/object type property
Semantics
class deletion/modification, invert property assertions, change class/property
hierarchy, assert class disjointness
Combinations of Variations
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 6 / 36
Benchmark: Gold Standard
The "correct answer sheet" used to judge the completeness and soundness of
the instance matching algorithms
Characteristics
Existence of errors / missing alignments
Representation: owl:sameAs and skos:exactMatch
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 7 / 36
Benchmark: Metrics
Precision P = tp
(tp+fn)
Recall R = tp
(tp+fp)
F-measure F = 2 × P × R
(P+R)
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 8 / 36
Instance Matching Benchmarks:
Desirable Attributes
Systematic Procedure matching tasks should be reproducible and the exe-
cution must be comparable
Availability benchmark should be available
Quality precise evaluation rules and high quality ontologies
must be provided
Equity evaluation process should not privilege any system
Dissemination benchmark should be used to evaluate instance
matching systems
Volume dataset size
Gold Standard gold standard should exist and be as accurate as pos-
sible
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 9 / 36
What about Benchmarks?
Instance matching techniques have, until recently, been
benchmarked in an ad-hoc way.
There is no standard way of benchmarking the performance
of the systems, when it comes to Linked Data.
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 10 / 36
Ontology Alignment Evaluation Initiative
IM benchmarks have been mainly driven forward by the Ontology
Alignment Evaluation Initiative (OAEI)
organizes annual campaign for ontology matching since 2005
hosts independent benchmarks
In 2009, OAEI introduced the Instance Matching (IM) Track
focuses on the evaluation of different instance matching techniques and tools
for Linked Data
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 11 / 36
Instance Matching Benchmarks
Bechmark Generators
Synthetic Benchmarks
Real Benchmarks
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 12 / 36
Semantic Web Instance Generation
(SWING) [FMN+11]
Semi automatic generator of Instance Matching Benchmarks
Contributed in the generation of IIMB Benchmarks of OAEI in 2010, 2011
and 2012 Instance Matching Tracks
Freely available at (https://code.google.com/p/swing-generator/)
All kind of variations supported into the benchmarks except multilinguality
Automatically produced gold standard
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 13 / 36
Lance [SDF+15b]
Flexible, generic and domain-independent benchmark generator which takes
into consideration RDFS and OWL constructs in order to evaluate instance
matching systems.
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 14 / 36
Lance [SDF+15b]
Lance provides support for:
Semantics-aware transformations
Complex class definitions (union, intersection)
Complex property definitions (functional properties, inverse functional
properties)
Disjointness (properties)
Standard value and structure based transformations
Weighted gold standard based on tensor factorization
Varying degrees of difficulty and fine-grained evaluation metrics
Available at http://github.com/jsaveta/Lance
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 15 / 36
Lance Architecture
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 16 / 36
Synthetic Benchmarks
Ontology Alignment Evaluation Benchmarks
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 17 / 36
Synthetic Instance Matching Benchmarks:
Overview (1)
IIMB
2009
IIMB
2010
PR
2010
IIMB
2011
Sandbox
2012
IIMB
2012
RDFT
2013
ID-
REC
2014
Author
Task
2015
Systematic Procedure
√ √ √ √ √ √ √ √ √
Availability
√ √ √ √ √
- -
√ √
Quality
√ √ √ √ √ √ √ √ √
Equity
√ √ √ √ √ √ √ √ √
Dissemination 6 3 6 1 3 4 4 5 5
Volume 0.2K 1.4K 0.86K 4K 0.375K 1.5K 0.43K 2.650K 10K
Gold Standard
√ √ √ √ √ √ √ √ √
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 18 / 36
Synthetic Instance Matching Benchmarks:
Overview (2)
IIMB
2009
IIMB
2010
PR
2010
IIMB
2011
Sandbox
2012
IIMB
2012
RDFT
2013
ID-
REC
2014
Author
Task
2015
Value Variations
√ √ √ √ √ √ √ √ √
Structural Variations
√ √ √ √
- - - + +
Logical Variations
√ √
-
√
-
√
- - -
Multilinguality - - - - - -
√ √ √
IIMB
2009
IIMB
2010
PR
2010
IIMB
2011
Sandbox
2012
IIMB
2012
RDFT
2013
ID-
REC
2014
Author
Task
2015
Blind Evaluations - - - - - -
√ √ √
1-n Mappings - -
√
- - -
√ √
-
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 19 / 36
Synthetic Instance Matching Benchmarks:
Overview (3)
IIMB
2009
IIMB
2010
PR
2010
IIMB
2011
Sandbox
2012
IIMB
2012
RDFT
2013
ID-
REC
2014
Author
Task
2015
Lance
2015
Systematic
Procedure
√ √ √ √ √ √ √ √ √ √
Availability
√ √ √ √ √
- -
√ √ √
Quality
√ √ √ √ √ √ √ √ √ √
Equity
√ √ √ √ √ √ √ √ √ √
Dissemination 6 3 6 1 3 4 4 5 5 2
Volume 0.2K 1.4K 0.86K 4K 0.375K 1.5K 0.43K 2.650K 10K > 1M
Gold
Standard
√ √ √ √ √ √ √ √ √ √
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 20 / 36
Synthetic Instance Matching Benchmarks:
Overview (4)
IIMB
2009
IIMB
2010
PR
2010
IIMB
2011
Sandbox
2012
IIMB
2012
RDFT
2013
ID-
REC
2014
Author
Task
2015
Lance
2015
Value
Variations
√ √ √ √ √ √ √ √ √ √
Structural
Variations
√ √ √ √
- - - + + +
Logical
Variations
√ √
-
√
-
√
- - - +
Multilinguality - - - - - -
√ √ √ √
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 21 / 36
Synthetic Instance Matching Benchmarks:
Overview (5)
IIMB
2009
IIMB
2010
PR
2010
IIMB
2011
Sandbox
2012
IIMB
2012
RDFT
2013
ID-
REC
2014
Author
Task
2015
Lance
2015
Blind
Evaluations
- - - - - -
√ √ √ √
1-n
Mappings
- -
√
- - -
√ √
- -
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 22 / 36
Real Benchmarks
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 23 / 36
Real Instance Matching Benchmarks:
Overview (1)
ARS DI 2010 DI 2011
Systematic Procedure
√ √ √
Availability
√ √
-
Quality
√ √ √
Equity
√ √ √
Dissemination 5 2 3
Volume 100K 6K NA
Gold Standard
√ √
+
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 24 / 36
Real Instance Matching Benchmarks:
Overview (2)
ARS DI 2010 DI 2011
Value Variations
√ √ √
Structural Variations
√ √
-
Logical Variations - - -
Multilinguality - - -
Blind Evaluations - - -
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 25 / 36
Wrapping Up
Multilinguality
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 26 / 36
Wrapping Up
Value Variations
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 27 / 36
Wrapping Up
Structural Variations
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 28 / 36
Wrapping Up
Logical Variations
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 29 / 36
Wrapping Up
Combinations of Variations
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 30 / 36
Wrapping Up
Scalability
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 31 / 36
Open Issues
Only one benchmark that tackles both, combination of variations and
scalability issues
Not enough IM benchmark using the full expressiveness of RDF/OWL
language
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 32 / 36
Systems
Systems can handle the value variations, the structural variation, and the
simple logical variations separately.
More work needed for complex variations (combination of value, structural,
and logical)
More work needed for structural variations
Enhancement of systems to cope with the clustering of the mappings (1-n
mappings)
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 33 / 36
Conclusions
Many instance matching benchmarks have been proposed
Each of them answering to some of the needs of instance matching systems.
It is essential to start creating benchmarks that will “show the way to the
future”
Extend the limits of existing systems.
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 34 / 36
Acknowledgment
This work was supported by grants from the EU H2020 Framework Programme
provided for the project HOBBIT (GA no. 688227).
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 35 / 36
References I
Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 36 / 36

More Related Content

What's hot (20)

PDF
Knowledge extraction in Web media: at the frontier of NLP, Machine Learning a...
Julien PLU
 
PDF
Enhancing Entity Linking by Combining NER Models
Julien PLU
 
PDF
Interactive Knowledge Discovery over Web of Data.
Mehwish Alam
 
PDF
Learning Commonalities in RDF
Sara EL HASSAD
 
PDF
Applications of Word Vectors in Text Retrieval and Classification
shakimov
 
PDF
Detecting paraphrases using recursive autoencoders
Feynman Liang
 
PDF
Learning Multilingual Semantic Parsers for Question Answering over Linked Dat...
shakimov
 
PDF
Recursive Autoencoders for Paraphrase Detection (Socher et al)
Feynman Liang
 
PDF
ESWC 2013 Poster: Representing and Querying Negative Knowledge in RDF
Fariz Darari
 
PDF
Reflection and Metadata
Michal Píše
 
PDF
Joint Word and Entity Embeddings for Entity Retrieval from Knowledge Graph
FedorNikolaev
 
PPT
F# and the DLR
Richard Minerich
 
PDF
Multiple Dispatch
Michal Píše
 
PDF
MediaEval 2015 - GTM-UVigo Systems for the Query-by-Example Search on Speech ...
multimediaeval
 
PDF
DCU Search Runs at MediaEval 2014 Search and Hyperlinking
multimediaeval
 
PDF
Federation and Navigation in SPARQL 1.1
net2-project
 
PPT
SPARQL in a nutshell
Fabien Gandon
 
PPTX
Exchange and Consumption of Huge RDF Data
Mario Arias
 
PDF
The Rise of Approximate Ontology Reasoning: Is It Mainstream Yet? --- Revisit...
Jeff Z. Pan
 
PPTX
Compact Representation of Large RDF Data Sets for Publishing and Exchange
WU (Vienna University of Economics and Business)
 
Knowledge extraction in Web media: at the frontier of NLP, Machine Learning a...
Julien PLU
 
Enhancing Entity Linking by Combining NER Models
Julien PLU
 
Interactive Knowledge Discovery over Web of Data.
Mehwish Alam
 
Learning Commonalities in RDF
Sara EL HASSAD
 
Applications of Word Vectors in Text Retrieval and Classification
shakimov
 
Detecting paraphrases using recursive autoencoders
Feynman Liang
 
Learning Multilingual Semantic Parsers for Question Answering over Linked Dat...
shakimov
 
Recursive Autoencoders for Paraphrase Detection (Socher et al)
Feynman Liang
 
ESWC 2013 Poster: Representing and Querying Negative Knowledge in RDF
Fariz Darari
 
Reflection and Metadata
Michal Píše
 
Joint Word and Entity Embeddings for Entity Retrieval from Knowledge Graph
FedorNikolaev
 
F# and the DLR
Richard Minerich
 
Multiple Dispatch
Michal Píše
 
MediaEval 2015 - GTM-UVigo Systems for the Query-by-Example Search on Speech ...
multimediaeval
 
DCU Search Runs at MediaEval 2014 Search and Hyperlinking
multimediaeval
 
Federation and Navigation in SPARQL 1.1
net2-project
 
SPARQL in a nutshell
Fabien Gandon
 
Exchange and Consumption of Huge RDF Data
Mario Arias
 
The Rise of Approximate Ontology Reasoning: Is It Mainstream Yet? --- Revisit...
Jeff Z. Pan
 
Compact Representation of Large RDF Data Sets for Publishing and Exchange
WU (Vienna University of Economics and Business)
 

Similar to Link Discovery Tutorial Part III: Benchmarking for Instance Matching Systems (20)

PDF
ISWC 2014 Tutorial - Instance Matching Benchmarks for Linked Data
Evangelia Daskalaki
 
PDF
Instance Matching Benchmarks for Linked Data - ESWC 2016 Tutorial
Holistic Benchmarking of Big Linked Data
 
PDF
Instance Matching Benchmarks in the ERA of Linked Data - ISWC2017
Holistic Benchmarking of Big Linked Data
 
PPTX
SPIMBENCH: A Scalable, Schema-Aware Instance Matching Benchmark for the Seman...
Graph-TA
 
PPTX
SPIMBENCH: A scalable, Schema-Aware Instance Matching Benchmark for the Seman...
Ioan Toma
 
PPTX
SPIMBENCH: A scalable, Schema-Aware Instance Matching Benchmark for the Seman...
LDBC council
 
PDF
BDVe Webinar Series: DataBench – Benchmarking Big Data. Arne Berre. Tue, Oct ...
Big Data Value Association
 
PPTX
Big Data Technical Benchmarking, Arne Berre, BDVe Webinar series, 09/10/2018
DataBench
 
PDF
Towards a Macrobenchmark Framework for Performance Analysis of Java Applications
Gábor Szárnyas
 
PDF
Hobbit presentation at Apache Big Data Europe 2016
Holistic Benchmarking of Big Linked Data
 
PDF
Benchmarks for Digital Preservation tools. Kresimir Duretec, Artur Kulmukhame...
12th International Conference on Digital Preservation (iPRES 2015)
 
PPTX
Improving Business Performance Through Big Data Benchmarking, Todor Ivanov, B...
DataBench
 
PDF
The value of benchmarking IT projects - H.S. van Heeringen
Harold van Heeringen
 
PPTX
Benchmarking for Big Data Applications with the DataBench Framework, Arne Ber...
DataBench
 
ODP
FOSDEM 2014: Social Network Benchmark (SNB) Graph Generator
LDBC council
 
PDF
2016 VLDB - The iBench Integration Metadata Generator
Boris Glavic
 
PDF
Benchmarks
Amit Kumar Rathi
 
PDF
Benchmarking Versioning for Big Linked Data
Graph-TA
 
PDF
Presentation of HOBBIT's versioning benchmark at Graph-TA
Holistic Benchmarking of Big Linked Data
 
PDF
Bytewise approximate matching, searching and clustering
Liwei Ren任力偉
 
ISWC 2014 Tutorial - Instance Matching Benchmarks for Linked Data
Evangelia Daskalaki
 
Instance Matching Benchmarks for Linked Data - ESWC 2016 Tutorial
Holistic Benchmarking of Big Linked Data
 
Instance Matching Benchmarks in the ERA of Linked Data - ISWC2017
Holistic Benchmarking of Big Linked Data
 
SPIMBENCH: A Scalable, Schema-Aware Instance Matching Benchmark for the Seman...
Graph-TA
 
SPIMBENCH: A scalable, Schema-Aware Instance Matching Benchmark for the Seman...
Ioan Toma
 
SPIMBENCH: A scalable, Schema-Aware Instance Matching Benchmark for the Seman...
LDBC council
 
BDVe Webinar Series: DataBench – Benchmarking Big Data. Arne Berre. Tue, Oct ...
Big Data Value Association
 
Big Data Technical Benchmarking, Arne Berre, BDVe Webinar series, 09/10/2018
DataBench
 
Towards a Macrobenchmark Framework for Performance Analysis of Java Applications
Gábor Szárnyas
 
Hobbit presentation at Apache Big Data Europe 2016
Holistic Benchmarking of Big Linked Data
 
Benchmarks for Digital Preservation tools. Kresimir Duretec, Artur Kulmukhame...
12th International Conference on Digital Preservation (iPRES 2015)
 
Improving Business Performance Through Big Data Benchmarking, Todor Ivanov, B...
DataBench
 
The value of benchmarking IT projects - H.S. van Heeringen
Harold van Heeringen
 
Benchmarking for Big Data Applications with the DataBench Framework, Arne Ber...
DataBench
 
FOSDEM 2014: Social Network Benchmark (SNB) Graph Generator
LDBC council
 
2016 VLDB - The iBench Integration Metadata Generator
Boris Glavic
 
Benchmarks
Amit Kumar Rathi
 
Benchmarking Versioning for Big Linked Data
Graph-TA
 
Presentation of HOBBIT's versioning benchmark at Graph-TA
Holistic Benchmarking of Big Linked Data
 
Bytewise approximate matching, searching and clustering
Liwei Ren任力偉
 
Ad

More from Holistic Benchmarking of Big Linked Data (20)

PDF
EARL: Joint Entity and Relation Linking for Question Answering over Knowledge...
Holistic Benchmarking of Big Linked Data
 
PDF
Benchmarking Big Linked Data: The case of the HOBBIT Project
Holistic Benchmarking of Big Linked Data
 
PDF
Assessing Linked Data Versioning Systems: The Semantic Publishing Versioning ...
Holistic Benchmarking of Big Linked Data
 
PDF
The DEBS Grand Challenge 2018
Holistic Benchmarking of Big Linked Data
 
PPTX
Benchmarking of distributed linked data streaming systems
Holistic Benchmarking of Big Linked Data
 
PDF
SQCFramework: SPARQL Query Containment Benchmarks Generation Framework
Holistic Benchmarking of Big Linked Data
 
PDF
LargeRDFBench: A billion triples benchmark for SPARQL endpoint federation
Holistic Benchmarking of Big Linked Data
 
PPTX
The DEBS Grand Challenge 2017
Holistic Benchmarking of Big Linked Data
 
PDF
4th Natural Language Interface over the Web of Data (NLIWoD) workshop and QAL...
Holistic Benchmarking of Big Linked Data
 
PDF
Scalable Link Discovery for Modern Data-Driven Applications (poster)
Holistic Benchmarking of Big Linked Data
 
PDF
Scalable Link Discovery for Modern Data-Driven Applications
Holistic Benchmarking of Big Linked Data
 
PDF
Extending LargeRDFBench for Multi-Source Data at Scale for SPARQL Endpoint F...
Holistic Benchmarking of Big Linked Data
 
PPTX
SPgen: A Benchmark Generator for Spatial Link Discovery Tools
Holistic Benchmarking of Big Linked Data
 
PDF
Introducing the HOBBIT platform into the Ontology Alignment Evaluation Campaign
Holistic Benchmarking of Big Linked Data
 
PDF
OKE2018 Challenge @ ESWC2018
Holistic Benchmarking of Big Linked Data
 
PDF
MOCHA 2018 Challenge @ ESWC2018
Holistic Benchmarking of Big Linked Data
 
PDF
Dynamic planning for link discovery - ESWC 2018
Holistic Benchmarking of Big Linked Data
 
PDF
Hobbit project overview presented at EBDVF 2017
Holistic Benchmarking of Big Linked Data
 
PDF
Leopard ISWC Semantic Web Challenge 2017 (poster)
Holistic Benchmarking of Big Linked Data
 
PDF
Leopard ISWC Semantic Web Challenge 2017
Holistic Benchmarking of Big Linked Data
 
EARL: Joint Entity and Relation Linking for Question Answering over Knowledge...
Holistic Benchmarking of Big Linked Data
 
Benchmarking Big Linked Data: The case of the HOBBIT Project
Holistic Benchmarking of Big Linked Data
 
Assessing Linked Data Versioning Systems: The Semantic Publishing Versioning ...
Holistic Benchmarking of Big Linked Data
 
The DEBS Grand Challenge 2018
Holistic Benchmarking of Big Linked Data
 
Benchmarking of distributed linked data streaming systems
Holistic Benchmarking of Big Linked Data
 
SQCFramework: SPARQL Query Containment Benchmarks Generation Framework
Holistic Benchmarking of Big Linked Data
 
LargeRDFBench: A billion triples benchmark for SPARQL endpoint federation
Holistic Benchmarking of Big Linked Data
 
The DEBS Grand Challenge 2017
Holistic Benchmarking of Big Linked Data
 
4th Natural Language Interface over the Web of Data (NLIWoD) workshop and QAL...
Holistic Benchmarking of Big Linked Data
 
Scalable Link Discovery for Modern Data-Driven Applications (poster)
Holistic Benchmarking of Big Linked Data
 
Scalable Link Discovery for Modern Data-Driven Applications
Holistic Benchmarking of Big Linked Data
 
Extending LargeRDFBench for Multi-Source Data at Scale for SPARQL Endpoint F...
Holistic Benchmarking of Big Linked Data
 
SPgen: A Benchmark Generator for Spatial Link Discovery Tools
Holistic Benchmarking of Big Linked Data
 
Introducing the HOBBIT platform into the Ontology Alignment Evaluation Campaign
Holistic Benchmarking of Big Linked Data
 
OKE2018 Challenge @ ESWC2018
Holistic Benchmarking of Big Linked Data
 
MOCHA 2018 Challenge @ ESWC2018
Holistic Benchmarking of Big Linked Data
 
Dynamic planning for link discovery - ESWC 2018
Holistic Benchmarking of Big Linked Data
 
Hobbit project overview presented at EBDVF 2017
Holistic Benchmarking of Big Linked Data
 
Leopard ISWC Semantic Web Challenge 2017 (poster)
Holistic Benchmarking of Big Linked Data
 
Leopard ISWC Semantic Web Challenge 2017
Holistic Benchmarking of Big Linked Data
 
Ad

Recently uploaded (20)

PPTX
Neuroinflammation and microglial subtypes
KanakChaudhary10
 
PPTX
Q1_Science 8_Week3-Day 1.pptx science lesson
AizaRazonado
 
PPTX
ION EXCHANGE CHROMATOGRAPHY NEW PPT (JA).pptx
adhagalejotshna
 
PPT
Restriction digestion of DNA for students of undergraduate and post graduate ...
DrMukeshRameshPimpli
 
PDF
Rapid protoplanet formation in the outer Solar System recorded in a dunite fr...
Sérgio Sacani
 
PDF
High-speedBouldersandtheDebrisFieldinDARTEjecta
Sérgio Sacani
 
DOCX
Paper - Taboo Language (Makalah Presentasi)
Sahmiral Amri Rajagukguk
 
PPTX
Cerebellum_ Parts_Structure_Function.pptx
muralinath2
 
PDF
Preserving brand authenticity amid AI-driven misinformation: Sustaining consu...
Selcen Ozturkcan
 
PDF
Unit-3 ppt.pdf organic chemistry - 3 unit 3
visionshukla007
 
PPTX
770043401-q1-Ppt-pe-and-Health-7-week-1-lesson-1.pptx
AizaRazonado
 
DOCX
Critical Book Review (CBR) - "Hate Speech: Linguistic Perspectives"
Sahmiral Amri Rajagukguk
 
PDF
Adding Geochemistry To Understand Recharge Areas - Kinney County, Texas - Jim...
Texas Alliance of Groundwater Districts
 
PPTX
Basal_ganglia_Structure_Function_Importance
muralinath2
 
DOCX
Paper - Suprasegmental Features (Makalah Presentasi)
Sahmiral Amri Rajagukguk
 
PPT
Experimental Design by Cary Willard v3.ppt
MohammadRezaNirooman1
 
PPTX
Class12_Physics_Chapter2 electric potential and capacitance.pptx
mgmahati1234
 
PDF
Portable Hyperspectral Imaging (pHI) for the enhanced recording of archaeolog...
crabbn
 
PDF
Global Congress on Forensic Science and Research
infoforensicscience2
 
PDF
A High-Caliber View of the Bullet Cluster through JWST Strong and Weak Lensin...
Sérgio Sacani
 
Neuroinflammation and microglial subtypes
KanakChaudhary10
 
Q1_Science 8_Week3-Day 1.pptx science lesson
AizaRazonado
 
ION EXCHANGE CHROMATOGRAPHY NEW PPT (JA).pptx
adhagalejotshna
 
Restriction digestion of DNA for students of undergraduate and post graduate ...
DrMukeshRameshPimpli
 
Rapid protoplanet formation in the outer Solar System recorded in a dunite fr...
Sérgio Sacani
 
High-speedBouldersandtheDebrisFieldinDARTEjecta
Sérgio Sacani
 
Paper - Taboo Language (Makalah Presentasi)
Sahmiral Amri Rajagukguk
 
Cerebellum_ Parts_Structure_Function.pptx
muralinath2
 
Preserving brand authenticity amid AI-driven misinformation: Sustaining consu...
Selcen Ozturkcan
 
Unit-3 ppt.pdf organic chemistry - 3 unit 3
visionshukla007
 
770043401-q1-Ppt-pe-and-Health-7-week-1-lesson-1.pptx
AizaRazonado
 
Critical Book Review (CBR) - "Hate Speech: Linguistic Perspectives"
Sahmiral Amri Rajagukguk
 
Adding Geochemistry To Understand Recharge Areas - Kinney County, Texas - Jim...
Texas Alliance of Groundwater Districts
 
Basal_ganglia_Structure_Function_Importance
muralinath2
 
Paper - Suprasegmental Features (Makalah Presentasi)
Sahmiral Amri Rajagukguk
 
Experimental Design by Cary Willard v3.ppt
MohammadRezaNirooman1
 
Class12_Physics_Chapter2 electric potential and capacitance.pptx
mgmahati1234
 
Portable Hyperspectral Imaging (pHI) for the enhanced recording of archaeolog...
crabbn
 
Global Congress on Forensic Science and Research
infoforensicscience2
 
A High-Caliber View of the Bullet Cluster through JWST Strong and Weak Lensin...
Sérgio Sacani
 

Link Discovery Tutorial Part III: Benchmarking for Instance Matching Systems

  • 1. Link Discovery Tutorial Benchmarking for Instance Matching Systems Axel-Cyrille Ngonga Ngomo(1) , Irini Fundulaki(2) , Mohamed Ahmed Sherif(1) (1) Institute for Applied Informatics, Germany (2) FORTH, Greece October 18th, 2016 Kobe, Japan Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 1 / 36
  • 2. The Questions(s) Instance matching research has led to the development of various systems. What are the problems that I wish to solve? What are the relevant key performance indicators? What is the behavior of the existing engines w.r.t. the key performance indicators? Which are the tool(s) that I should use for my data and for my use case? Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 2 / 36
  • 3. Importance of Benchmarking Benchmarks exist To allow adequate measurements of systems To provide evaluation of engines for real (or close to real) use cases Provide help Designers and Developers to assess the performance of their tools Users to compare the different available tools and evaluate suitability for their needs Researchers to compare their work to others Leads to improvements: Vendors can improve their technology Researchers can address new challenges Current benchmark design can be improved to cover new necessities and application domains Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 3 / 36
  • 4. The Answer: Benchmark your engines! Instance Matching/Linking Benchmark comprises of Datasets: The raw material of the benchmarks. These are the source and the target dataset that will be matched together to find the links between resources Test Cases: Address heterogeneities (structural, value, semantic) of the datasets to be matched Gold Standard (Ground Truth / Reference Alignment): The "correct answer sheet" used to judge the completeness and soundness of the instance matching algorithms Metrics: The performance metric(s) that determine the systems behaviour and performance Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 4 / 36
  • 5. Benchmark Datasets: Characteristics Nature Real Datasets: Widely used datasets from a domain of interest + Realistic conditions for heterogeneity problems + Realistic distributions - Error prone, hard to create Reference Alignment Synthetic Datasets: Produced with a data generator (that hopefully produces data with interesting characteristics + Fully controlled test conditions + Accurate, Easy to create Reference Alignments - Unrealistic distributions - Systematic heterogeneity problems Schema Datasets to be matched have the same or different schemas Domain Datasets come from the same or different domains Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 5 / 36
  • 6. Benchmark Test Cases: Variations Value Name style abbreviations, Typographical errors, change format (date/gender/number), synonym change, language change (multilinguality) Structural Change property depth, Delete/add property, split property values, transformation of object/data to data/object type property Semantics class deletion/modification, invert property assertions, change class/property hierarchy, assert class disjointness Combinations of Variations Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 6 / 36
  • 7. Benchmark: Gold Standard The "correct answer sheet" used to judge the completeness and soundness of the instance matching algorithms Characteristics Existence of errors / missing alignments Representation: owl:sameAs and skos:exactMatch Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 7 / 36
  • 8. Benchmark: Metrics Precision P = tp (tp+fn) Recall R = tp (tp+fp) F-measure F = 2 × P × R (P+R) Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 8 / 36
  • 9. Instance Matching Benchmarks: Desirable Attributes Systematic Procedure matching tasks should be reproducible and the exe- cution must be comparable Availability benchmark should be available Quality precise evaluation rules and high quality ontologies must be provided Equity evaluation process should not privilege any system Dissemination benchmark should be used to evaluate instance matching systems Volume dataset size Gold Standard gold standard should exist and be as accurate as pos- sible Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 9 / 36
  • 10. What about Benchmarks? Instance matching techniques have, until recently, been benchmarked in an ad-hoc way. There is no standard way of benchmarking the performance of the systems, when it comes to Linked Data. Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 10 / 36
  • 11. Ontology Alignment Evaluation Initiative IM benchmarks have been mainly driven forward by the Ontology Alignment Evaluation Initiative (OAEI) organizes annual campaign for ontology matching since 2005 hosts independent benchmarks In 2009, OAEI introduced the Instance Matching (IM) Track focuses on the evaluation of different instance matching techniques and tools for Linked Data Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 11 / 36
  • 12. Instance Matching Benchmarks Bechmark Generators Synthetic Benchmarks Real Benchmarks Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 12 / 36
  • 13. Semantic Web Instance Generation (SWING) [FMN+11] Semi automatic generator of Instance Matching Benchmarks Contributed in the generation of IIMB Benchmarks of OAEI in 2010, 2011 and 2012 Instance Matching Tracks Freely available at (https://code.google.com/p/swing-generator/) All kind of variations supported into the benchmarks except multilinguality Automatically produced gold standard Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 13 / 36
  • 14. Lance [SDF+15b] Flexible, generic and domain-independent benchmark generator which takes into consideration RDFS and OWL constructs in order to evaluate instance matching systems. Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 14 / 36
  • 15. Lance [SDF+15b] Lance provides support for: Semantics-aware transformations Complex class definitions (union, intersection) Complex property definitions (functional properties, inverse functional properties) Disjointness (properties) Standard value and structure based transformations Weighted gold standard based on tensor factorization Varying degrees of difficulty and fine-grained evaluation metrics Available at http://github.com/jsaveta/Lance Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 15 / 36
  • 16. Lance Architecture Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 16 / 36
  • 17. Synthetic Benchmarks Ontology Alignment Evaluation Benchmarks Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 17 / 36
  • 18. Synthetic Instance Matching Benchmarks: Overview (1) IIMB 2009 IIMB 2010 PR 2010 IIMB 2011 Sandbox 2012 IIMB 2012 RDFT 2013 ID- REC 2014 Author Task 2015 Systematic Procedure √ √ √ √ √ √ √ √ √ Availability √ √ √ √ √ - - √ √ Quality √ √ √ √ √ √ √ √ √ Equity √ √ √ √ √ √ √ √ √ Dissemination 6 3 6 1 3 4 4 5 5 Volume 0.2K 1.4K 0.86K 4K 0.375K 1.5K 0.43K 2.650K 10K Gold Standard √ √ √ √ √ √ √ √ √ Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 18 / 36
  • 19. Synthetic Instance Matching Benchmarks: Overview (2) IIMB 2009 IIMB 2010 PR 2010 IIMB 2011 Sandbox 2012 IIMB 2012 RDFT 2013 ID- REC 2014 Author Task 2015 Value Variations √ √ √ √ √ √ √ √ √ Structural Variations √ √ √ √ - - - + + Logical Variations √ √ - √ - √ - - - Multilinguality - - - - - - √ √ √ IIMB 2009 IIMB 2010 PR 2010 IIMB 2011 Sandbox 2012 IIMB 2012 RDFT 2013 ID- REC 2014 Author Task 2015 Blind Evaluations - - - - - - √ √ √ 1-n Mappings - - √ - - - √ √ - Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 19 / 36
  • 20. Synthetic Instance Matching Benchmarks: Overview (3) IIMB 2009 IIMB 2010 PR 2010 IIMB 2011 Sandbox 2012 IIMB 2012 RDFT 2013 ID- REC 2014 Author Task 2015 Lance 2015 Systematic Procedure √ √ √ √ √ √ √ √ √ √ Availability √ √ √ √ √ - - √ √ √ Quality √ √ √ √ √ √ √ √ √ √ Equity √ √ √ √ √ √ √ √ √ √ Dissemination 6 3 6 1 3 4 4 5 5 2 Volume 0.2K 1.4K 0.86K 4K 0.375K 1.5K 0.43K 2.650K 10K > 1M Gold Standard √ √ √ √ √ √ √ √ √ √ Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 20 / 36
  • 21. Synthetic Instance Matching Benchmarks: Overview (4) IIMB 2009 IIMB 2010 PR 2010 IIMB 2011 Sandbox 2012 IIMB 2012 RDFT 2013 ID- REC 2014 Author Task 2015 Lance 2015 Value Variations √ √ √ √ √ √ √ √ √ √ Structural Variations √ √ √ √ - - - + + + Logical Variations √ √ - √ - √ - - - + Multilinguality - - - - - - √ √ √ √ Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 21 / 36
  • 22. Synthetic Instance Matching Benchmarks: Overview (5) IIMB 2009 IIMB 2010 PR 2010 IIMB 2011 Sandbox 2012 IIMB 2012 RDFT 2013 ID- REC 2014 Author Task 2015 Lance 2015 Blind Evaluations - - - - - - √ √ √ √ 1-n Mappings - - √ - - - √ √ - - Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 22 / 36
  • 23. Real Benchmarks Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 23 / 36
  • 24. Real Instance Matching Benchmarks: Overview (1) ARS DI 2010 DI 2011 Systematic Procedure √ √ √ Availability √ √ - Quality √ √ √ Equity √ √ √ Dissemination 5 2 3 Volume 100K 6K NA Gold Standard √ √ + Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 24 / 36
  • 25. Real Instance Matching Benchmarks: Overview (2) ARS DI 2010 DI 2011 Value Variations √ √ √ Structural Variations √ √ - Logical Variations - - - Multilinguality - - - Blind Evaluations - - - Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 25 / 36
  • 26. Wrapping Up Multilinguality Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 26 / 36
  • 27. Wrapping Up Value Variations Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 27 / 36
  • 28. Wrapping Up Structural Variations Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 28 / 36
  • 29. Wrapping Up Logical Variations Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 29 / 36
  • 30. Wrapping Up Combinations of Variations Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 30 / 36
  • 31. Wrapping Up Scalability Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 31 / 36
  • 32. Open Issues Only one benchmark that tackles both, combination of variations and scalability issues Not enough IM benchmark using the full expressiveness of RDF/OWL language Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 32 / 36
  • 33. Systems Systems can handle the value variations, the structural variation, and the simple logical variations separately. More work needed for complex variations (combination of value, structural, and logical) More work needed for structural variations Enhancement of systems to cope with the clustering of the mappings (1-n mappings) Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 33 / 36
  • 34. Conclusions Many instance matching benchmarks have been proposed Each of them answering to some of the needs of instance matching systems. It is essential to start creating benchmarks that will “show the way to the future” Extend the limits of existing systems. Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 34 / 36
  • 35. Acknowledgment This work was supported by grants from the EU H2020 Framework Programme provided for the project HOBBIT (GA no. 688227). Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 35 / 36
  • 36. References I Ngonga Ngomo et al. (AKSW & FORTH) LD Tutorial:Benchmarking October 17, 2016 36 / 36