Complex track - Taxon dataset - Evaluation

The output alignments can be downloaded here.

Evaluation measures

Differently from the last years, in 2021 we manually evaluate the quality of the generated alignments, in terms of precision.

Execution environment

The alignment systems have been executed on a Ubuntu Linux machine configured with 32GB of RAM running under a Intel Core CPU 2.00GHz x8 cores. All measurements are based on a single run.

Participants

Only two systems applied for the complex track: AMLC and AMD, among which only AMLC was able to produce results. AMD was not able to parse the files, hence it produced no result.

The track manager launched all the systems (not only those that applied for the complex track) using the complex track dataset. Three experts evaluated separately the results of AMLC, Logmap and LogmapLite systems. Among them, one expert evaluated the results of all systems. It was the first time that those experts evaluated complex alignments using EDOAL language.

Evaluation results

Table below presents the synthesis of the manual evaluation produced by the three experts. Each cell contains the number of alignments and among them the number of validated alignments.

Alo2Vec AMD AMLC ATM Fine-TOM Gmap LogMap LogMapLite LSMatch TOM Wiki
AgronomicTaxon-agrovoc 0 0 0 0 0 0 1|0 [1 vocab] 5|3 [2 vocab trivial] 0 0 0
AgronomicTaxon-DBpedia 18|17 [8 vocab] 0 0 13|10 [1 vocab] 10|4 0 0 0 1|1 [1 vocab] 0 18|17 [8 vocab]
AgronomicTaxon-Taxref 3|2 [1 vocab] 0 0 18|18 [1 vocab] 1|1 [1 vocab trivial] 0 2|0 0 0 1|1 [1 vocab trivial] 2|2 [1 vocab trivial]
Agrovoc-agronomicTaxon 0 0 0 0 0 0 1|0 [1 vocab] 0 0 0 0
Agrovoc-DBpedia 0 0 0 0 0 0 5|4 0 0 0 0
Agrovoc-Taxref 0 0 3|1 [1 vocab trivial] 0 0 0 2|1 0 0 0 0
DBpedia-agronomicTaxon 18|17 [8 vocab] 0 0 13|7 11|1 0 0 0 1|1 [1 vocab] 018|17 [8 vocab]
DBpedia-agrovoc 0 0 0 0 0 0 5|4 0 0 0 0
DBpedia-Taxref 831 manually validated: 365|3 [3 vocab] 0 5|1 [1 vocab] 1033 manually validated: 45|43 39|22 [4 vocab trivial] 0 272 manually validated: 109|35 [2 vocab trivial] 0 1|1 3|3 [ 1 vocab + 2 vocab trivial] 830 validated manually: 205|2 [2 vocab]
Taxref-agronomicTaxon 3|2 [1 vocab] 0 0 18|18 [1 vocab] 6|5 [1 vocab trivial] 0 3|2 0 0 1|1 [1 vocab trivial] 2|2 [1 vocab trivial]
Taxref-agrovoc 0 0 3|1 [1 vocab trivial] 0 0 0 2|1 0 0 0 0
Taxref-DBpedia 831 manually validated: 95|0 0 4|0 1033 0 0 322 validated manually: 52|24 [2 vocab trivial] 0 1|1 0 830
# total of alignments 1704 0 15 2128 67 0 615 5 4 5 1700
# evaluated by at least one expert 502 0 15 107 67 0 182 5 4 5 245
# valid alignments 41 (8%) 0 3 (20%) 96 (89%) 33 (49%) 0 71 (39%) 3 (60%) 4 (100%) 5 (100%) 40 (16%)

Discussion

First remark, the experts expected to evaluate the links between taxonomic entities, i.e. taxa or scientific names, belonging to one of the proposed taxonomic registers (AgronomicTaxon, Agrovoc, DBpedia, TAXREF-LD). However, only little systems were able to find alignments between these entities. In many other cases, the aligned resources are terms from the vocabularies shared by several taxonomic registers (e.g. properties from SKOS, Dublin Core Terms or DOLCE+DnS Ultralite). This type of match is named “vocab” in table X. For instance, Logmap states that DOLCE+DnS Ultralite property hasPart (http://www.ontologydesignpatterns.org/ont/dul/DUL.owl#hasPart) is equivalent to Agrovoc property hasPart (http://aims.fao.org/aos/agrontology#hasPart) . Furthermore, some systems propose trivial alignments of vocabulary terms with themselves. For example, ALMC states that dcterm:created in Agrovoc is equivalent to dcterm:created in TAXREF-LD, which is true but useless. This type of trivial matches is named "vocab trivial" in table above.

Some complex alignments involve some SKOS properties related to creation or modification dates of a resource in the taxonomic register. This information is not relevant to identify entity alignments.

Some alignments involve entities belonging to external taxonomic registers (e.g. Geospecies or Umbold). The experts assume that some of these external entities come from some alignments that were declared in the taxonomic registers, e.g. TAXREF-LD comes with alignments with Geospecies.

A discussion between the experts emerged about this type of alignment and several questions came up:

Only the ATM, Fine-TOM and Logmap system seems capable of proposing valid alignments between taxa. Moreover, Logmap seems capable of dealing with challenge C2, that is, align two taxa that have different reference names but where the reference name on one side is considered a synonym on the other side.

To conclude, overall the experts were surprised that so little alignments were proposed. The taxonomic registers contain thousands to hundreds of thousands of taxa each, yet several tools were not able to propose more than a handful of alignments on vocabulary terms. No valid complex alignments were proposed between taxa. Only systems based on simple alignments were able to detect some of them.

Furthermore, the experts feel like a new interface is needed to visualize the results and ease the evaluation process. This interface should help filter out trivial matches and mention the origin of the entity referenced in the alignment (external resource, shared vocabulary). Moreover, the experts had difficulties understanding the EDOAL language. We expect that next year the evaluation process will integrate some tools meant to simplify the exploitation of EDOAL files like the edoal2csv tool [1]. Regarding taxonomic registers, a new resource about taxonomy was identified by the experts: the NCBI taxonomy available on the Agroportal repository. The TAXREF-LD taxonomic register used in the complex track is an old version. A new version is available (version 13) on the web. Unfortunately this graph is huge and could not be parsed by some system. A solution is to select a subset of TAXREF-LD covering only plant taxonomy. This entails the question about whether the other large taxonomic registers (such as NCBI or DBpedia) should also be limited to plant taxonomy.

References

[1] https://github.com/liseda-lab/EDOAL-2-CSV