There are three subtasks related to this evaluation:
Subtask 1 is evaluated based on standard precision, recall and F-measure, while subtasks 2 and 3 are evaluated using semantic precision, recall and F-measure, as defined in [1].
None of the alignment systems entered in this year's OAEI are capable of producing results for subtasks 2 and 3, so the table below shows the results of the systems that produced results on any of the ontology pairs within the test set as evaluated based on subtask 1. The line for ABC (Alignment By Comments) is a baseline matcher that simply considers an entity in the target ontology (TE) to be related to a source ontology entity (SE) if either the label or comment of SE mentions the label of TE. A "-" indicates that the system did not produce any results or did not run to completion for that particular ontology pair.
Matcher | Precision | Recall | F-measure |
---|---|---|---|
Hydro3-SWO | |||
ABC | 0.478 | 0.289 | 0.360 |
ALOD2Vec | 1.000 | 0.132 | 0.233 |
DOME | 1.000 | 0.132 | 0.233 |
FMapX | 0.833 | 0.237 | 0.369 |
KEPLER | 1.000 | 0.158 | 0.273 |
LogMap | 1.000 | 0.158 | 0.273 |
POMAP++ | 1.000 | 0.132 | 0.233 |
XMap | 0.833 | 0.237 | 0.369 |
HydrOntology(translated)-SWO | |||
ABC | 0.176 | 0.057 | 0.090 |
ALOD2Vec | - | - | - |
DOME | 0.263 | 0.024 | 0.044 |
FMapX | - | - | - |
KEPLER | - | - | - |
LogMap | 0.778 | 0.033 | 0.063 |
POMAP++ | 0.667 | 0.012 | 0.024 |
XMap | - | - | - |
HydrOntology(native)-SWO | |||
ABC | 0.250 | 0.005 | 0.010 |
ALOD2Vec | - | - | - |
DOME | 0.000 | 0.000 | 0.000 |
FMapX | - | - | - |
KEPLER | - | - | - |
LogMap | - | - | - |
POMAP++ | - | - | - |
XMap | - | - | - |
Cree-SWO | |||
ABC | 0.833 | 0.139 | 0.240 |
ALOD2Vec | 1.000 | 0.069 | 0.129 |
DOME | 0.119 | 0.069 | 0.087 |
FMapX | 1.000 | 0.042 | 0.081 |
KEPLER | 1.000 | 0.042 | 0.081 |
LogMap | - | - | - |
POMAP++ | - | - | - |
XMap | - | - | - |
The Hydro3-SWO ontology pair is the closest to existing OAEI benchmarks -- the majority of relationships between these two ontologies are 1-to-1 equivalences, so it is not surprising that the performance is best on this pair. The low performance on the native (Spanish) version of the HydrOntology-SWO pair highlights the lack of multilingual support in many existing matching systems. The Cree ontology also contains many non-English labels, but the comments are in English. It seems that most existing systems do not currently take advantage of this information, leading the baseline matcher to outperform the others in terms of F-measure. The recall of the baseline system is in general higher than the other matchers because many existing matchers filter their match candidates to produce 1-to-1 alignments whereas the alignments in this track are generally 1:n or m:n.
[1] Euzenat, Jérôme. "Semantic Precision and Recall for Ontology Alignment Evaluation." IJCAI. Vol. 7. 2007.