Ontology Alignment Evaluation Initiative - OAEI-2019 CampaignOAEI OAEI

Complex track - Taxon dataset - Evaluation

The output alignments can be downloaded here.

Evaluation measures

The main focus of the evaluation here is to be task-oriented. First, we manually evaluate the quality of the generated alignments, in terms of precision. Second, we evaluate the generated correspondences using a SPARQL query rewriting system and manually measure their ability of answering a set of queries over each dataset.

Precision

We excluded from the evaluation the generated correspondences implying identical IRIs such as owl:Class = owl:Class. As instance matching is not the focus of this evaluation, we do not consider instance correspondences. In order to measure the Precision of the output alignments, each correspondence (except if in the categories mentioned above) was manually verified and classified:

This gives 4 Precision scores:

The confidence of the correspondence is not taken into account. All the systems output only equivalence correspondences so the relation of the correspondence was not considered in the evaluation.

Coverage

6 reference equivalent queries per ontology are used to assess how an alignment covers them. The alignment is used to rewrite every reference. For each reference query, the best rewritten query is classified as equivalent, more specific, more general or overlapping.

The same scores as for the Precision are calculated for the Coverage:

Participants

We ran all the systems which applied for the Complex, Conference and Anatomy tracks: AGM, Alin, AML, AMLC, DOME, CANARD, FCAMap-KG, Lily, LogMap, LogMapBio, LogMapKG, LogMapLt, ONTMAT1, POMAP++, Wiktionary. Out of the 15 tested systems, 5 were not able to output alignments:

All the other systems (AGM, AML, CANARD, FCAMap-KG, LogMap, LogMapBio, LogMapLt, ONTMAT1, POMAP++) output at least an alignment. ONTMAT1 only output correspondences with identical URIs which were filtered for the evaluation. The only system to output complex correspondences was CANARD.

This year, every system was evaluated on their oriented alignment output, especially for the Coverage calculation. For example, let's take a set of equivalent correspondences: Q={SELECT ?x WHERE{ ?x a agtx:Taxon}, SELECT ?x WHERE{ ?x a dbo:Species}}. If an output alignment agronomicTaxon-dbpedia contains (agtx:Taxon,dbo:Species,≡) but the alignment dbpedia-agronomicTaxon does NOT contain (dbo:Species, agtx:Taxon, ≡). The coverage score of Q for the pair agronomicTaxon-dbpedia is 1 but the coverage score of Q for dbpedia-agronomicTaxon is 0. Last year the evaluation was non-oriented, so the coverage score of Q would be the same (1.0) for agronomicTaxon-dbpedia and dbpedia-agronomicTaxon.

Evaluation results

The systems have been executed on a Ubuntu 16.04 machine configured with 16GB of RAM running under a i7-4790K CPU 4.00GHz x 8 processors. All measurements are based on a single run.

Number of correspondences per type per system and runtime over the track.

"identical" shows the number of correspondences which align two identical URIs, instance, the number of correspondences which align two instances

tool (1:1) (1:n) (m:1) (m:n) identical instance total filtered runtime(s) runtime (min)
AGM 169 0 0 0 11 3029 3209 169 11593 193
AML 30 0 0 0 0 0 30 30 2444 41
CANARD 67 540 15 169 0 0 791 791 30712 512
FCAMap-KG 20 0 0 0 10 0 30 20 1284 21
LogMap 38 0 0 0 58 0 96 38 1544 26
LogMapBio 36 0 0 0 54 0 90 36 2009 33
LogMapKG 34 0 0 0 58 12776 12868 34 1654 28
LogMapLt 972 0 0 0 1528 7628 10128 972 1503 25
ONTMAT1 0 0 0 0 12 0 12 0 1586 26
POMAP++ 1 0 0 0 25 7 33 1 2252 38

Results over the Taxon dataset

Precision Coverage
tool classical recall oriented precision oriented overlap classical recall oriented precision oriented overlap
AGM 0.055 0.122 0.101 0.134 0.028 0.042 0.035 0.042
AML 0.533 0.533 0.533 0.533 0 0 0 0
CANARD 0.083 0.455 0.363 0.914 0.139 0.285 0.264 0.361
FCAMap-KG 0.625 0.875 0.875 0.958 0.028 0.049 0.049 0.056
LogMap 0.625 0.729 0.729 0.792 0.111 0.132 0.132 0.139
LogMapBio 0.537 0.659 0.659 0.723 0.083 0.104 0.104 0.111
LogMapKG 0.547 0.641 0.641 0.686 0.139 0.16 0.16 0.167
LogMapLt 0.221 0.365 0.365 0.413 0.083 0.135 0.135 0.153
POMAP++ 1 1 1 1 0 0 0 0

Discussion

LogMapLt and AGM have output correspondences such as agronomicTaxon:PhylumRank = dbo:phylum. As the source member is a class and the target member is an object property, these correspondences were counted as incorrect. However, they would be equivalent if transformed into complex correspondences such as (agronomicTaxon:PhylumRank, Exists dbo:phylum^-, ≡).

CANARD has the longest runtime by far. It took 8h32min to complete the Taxon matching task. It is the only matcher which outputs complex correspondences on this dataset.

The two system which cover the most queries by enabling an equivalent query output are LogMapKG and CANARD: their classical Coverage score is 0.139. CANARD enables to cover more queries with more specific, more general or overlapping answers.

In comparison with last year’s results, the Coverage score of CANARD, LogMap and LogMapBio have slightly improved (comparing only equivalent queries). This shows improvement especially in the case of oriented aligning systems such as CANARD as last year's evaluation was not oriented.

The classical Precision score of CANARD has decreased in comparison with last year's results while those of AML, LogMap, LogMapBio, LogMapLt, POMAP++ have increased.

POMAP++ output only one correspondence which was not filtered. This correspondence is correct, therefore, POMAP++ has the highest Precision scores 1.0. The correspondence could not be used to rewrite the reference queries, so its Coverage scores are 0.0.

With respect to the technical environment, the SEALS system as it is now was probably not adapted to deal with big knowledge bases as the loading phase got very slow. The use of SPARQL endpoints instead of two ontology files would make more sense given that many LOD repositories provide one.