Ontology Alignment Evaluation Initiative - OAEI-2021 Campaign OAEI OAEI

Results for OAEI 2021 - Knowledge Graph Track

Matching systems

As a pre-test, we executed all systems submitted to OAEI (even if they are not registered for the track) on a very small matching example (dataset) with a similar structure and shape like the real knowledge graphs (in fact, they are a small subset of them). It showed that not all matching systems are able to complete this small task due to exceptions or other failures. The following matching systems produced an exception: For AMD we adjusted the requirements.txt file to also include lxml which is necessary to execute the matcher. Thus, we executed the following systems: The source code for the baseline matchers is available. The baselineLabel matcher matches all resources which share the same rdfs:label. (in case multiple resources share the same label, all of them are matched). BaselineAltLabel is additionally using skos:altLabel. Again, in cases where multiple resources share a common label, all those resources are matched in a cross product manner.

Experimental setting

The evaluation is executed on a virtual machine(VM) with 32GB of RAM and 16 vCPUs (2.4 GHz). The operating system is debian 9 with openjdk version "1.8.0_265".

We used the MELT toolkit for the evaluation which internally uses the SEALS client (version 7.0.5) to execute matcher packaged with SEALS. Matching systems which use the web packaging, are executed with the MatcherHTTPCall class. The reported times includes the environment preparation of SEALS as well as the file upload to the docker container (the start of the container is not timed). The alignments were evaluated based on Precision, Recall and F-Measure for classes, properties and instances (each in isolation). Our partial gold standard consist of 1:1 mappings extracted from links contained in wiki pages (cross wiki links). The schema was matched by ontology experts. We assume that in each knowledge graph, only one representation of one concept exists. This means if we have the mapping in our gold standard we can count the mapping as a false positive (the assumption here is that in the seconds knowledge graph no similar concept to B exists). The value of false negatives is only increased if we have a 1:1 mapping and it is not found by a matcher. The source code for generating the evaluation results is also available.

We imposed a maximum execution time of 24h per task, however, that time limit was never exceeded.

Generated dashboard / CSV file

We also generated an online dashboard with the help of the MELT framework. Have a look at the knowledge graph results here (it may take some seconds to load due to 200 000 correspondences).
Moreover, we also generated a CSV file which allows to analyze each matcher on a correspondence level. This should help matcher developers to increase the matcher performance.

Alignment results

The generated alignment files are also available.

Results overview

class property instance overall
SystemTime#testcases SizePrec.F-m.Rec. SizePrec.F-m.Rec. SizePrec.F-m.Rec. SizePrec.F-m.Rec.
ALOD2Vec00:21:525 20.01.00 (1.00)0.80 (0.80)0.67 (0.67) 76.80.94 (0.94)0.95 (0.95)0.97 (0.97) 4893.40.91 (0.91)0.87 (0.87)0.83 (0.83) 4990.20.91 (0.91)0.87 (0.87)0.83 (0.83)
AMD00:37:472 23.00.40 (1.00)0.25 (0.62)0.18 (0.45) 0.00.00 (0.00)0.00 (0.00)0.00 (0.00) 0.00.00 (0.00)0.00 (0.00)0.00 (0.00) 23.00.40 (1.00)0.00 (0.00)0.00 (0.00)
AML00:50:265 23.60.98 (0.98)0.89 (0.89)0.81 (0.81) 48.40.92 (0.92)0.70 (0.70)0.57 (0.57) 6802.80.90 (0.90)0.85 (0.85)0.80 (0.80) 6874.80.90 (0.90)0.85 (0.85)0.80 (0.80)
ATMatcher00:19:345 25.60.97 (0.97)0.87 (0.87)0.79 (0.79) 78.80.97 (0.97)0.96 (0.96)0.95 (0.95) 4859.00.89 (0.89)0.85 (0.85)0.80 (0.80) 4963.40.89 (0.89)0.85 (0.85)0.81 (0.81)
BaselineAltLabel00:11:375 16.41.00 (1.00)0.74 (0.74)0.59 (0.59) 47.80.99 (0.99)0.79 (0.79)0.66 (0.66) 4674.80.89 (0.89)0.84 (0.84)0.80 (0.80) 4739.00.89 (0.89)0.84 (0.84)0.80 (0.80)
BaselineLabel00:11:275 16.41.00 (1.00)0.74 (0.74)0.59 (0.59) 47.80.99 (0.99)0.79 (0.79)0.66 (0.66) 3641.80.95 (0.95)0.81 (0.81)0.71 (0.71) 3706.00.95 (0.95)0.81 (0.81)0.71 (0.71)
Fine-TOM14:55:095 19.21.00 (1.00)0.80 (0.80)0.66 (0.66) 29.00.40 (0.40)0.39 (0.39)0.38 (0.38) 4116.00.92 (0.92)0.83 (0.83)0.76 (0.76) 4164.20.92 (0.92)0.83 (0.83)0.75 (0.75)
KGMatcher04:55:325 23.21.00 (1.00)0.79 (0.79)0.66 (0.66) 0.00.00 (0.00)0.00 (0.00)0.00 (0.00) 3789.60.94 (0.94)0.82 (0.82)0.74 (0.74) 3812.80.94 (0.94)0.82 (0.82)0.72 (0.72)
LogMap01:04:455 19.40.93 (0.93)0.81 (0.81)0.71 (0.71) 0.00.00 (0.00)0.00 (0.00)0.00 (0.00) 4012.40.90 (0.90)0.78 (0.78)0.69 (0.69) 4031.80.90 (0.90)0.77 (0.77)0.68 (0.68)
LSMatch02:02:555 18.41.00 (1.00)0.78 (0.78)0.64 (0.64) 0.00.00 (0.00)0.00 (0.00)0.00 (0.00) 0.00.00 (0.00)0.00 (0.00)0.00 (0.00) 18.41.00 (1.00)0.01 (0.01)0.00 (0.00)
OTMapOnto00:48:254 122.50.59 (0.73)0.61 (0.77)0.64 (0.80) 0.00.00 (0.00)0.00 (0.00)0.00 (0.00) 0.00.00 (0.00)0.00 (0.00)0.00 (0.00) 122.50.59 (0.73)0.01 (0.01)0.00 (0.01)
TOM23:30:255 19.41.00 (1.00)0.83 (0.83)0.71 (0.71) 0.00.00 (0.00)0.00 (0.00)0.00 (0.00) 311.40.91 (0.91)0.12 (0.12)0.06 (0.06) 330.80.92 (0.92)0.12 (0.12)0.06 (0.06)
Wiktionary00:43:185 22.01.00 (1.00)0.80 (0.80)0.67 (0.67) 79.80.94 (0.94)0.95 (0.95)0.97 (0.97) 4894.40.91 (0.91)0.87 (0.87)0.83 (0.83) 4996.20.91 (0.91)0.87 (0.87)0.83 (0.83)
Aggregated results per matcher, divided into class, property, instance, and overall alignments. Time is displayed as HH:MM:SS. Column #testcases indicates the number of testcases where the tool is able to generate (non empty) alignments. Column size indicates the averaged number of system correspondences. Two kinds of results are reported: (1) those not distinguishing empty and erroneous (or not generated) alignments, and (2) those considering only non empty alignments (value between parenthesis).
AMD and OTMapOnto could not return results for all test cases. Furthermore AMD, LSMatch, and OTMapOnto only return class correspondences. KGMatcher and TOM do not return property alignments. This is propbably due to the fact that they don't care about rdf:property. The longest runtime was used by TOM with 23 hours and nearly 31 minutes for all 5 test cases.

Test case specific results

Overall results

This result table shows the overall performance (without dividing into class, property or instance) of the matchers.
marvelcinematicuniverse-marvel memoryalpha-memorybeta memoryalpha-stexpanded starwars-swg starwars-swtor
SizePrec.F-m.Rec. SizePrec.F-m.Rec. SizePrec.F-m.Rec. SizePrec.F-m.Rec. SizePrec.F-m.Rec.
ALOD2Vec 30930.860.760.68 134550.920.910.90 34020.920.920.93 21850.920.830.75 28160.930.920.92
AMD 00.000.000.00 211.000.000.00 251.000.010.00 00.000.000.00 00.000.000.00
AML 46870.850.680.56 184390.910.890.87 37950.930.930.92 35150.900.810.74 39380.930.910.90
ATMatcher 35160.670.590.53 130060.960.930.91 32810.960.940.92 22430.930.840.76 27710.950.930.91
BaselineAltLabel 25740.860.760.68 135140.880.890.89 32300.880.900.92 17120.920.740.63 26650.920.910.90
BaselineLabel 18790.900.690.56 105520.950.850.77 25820.980.900.83 12450.960.680.53 22720.950.890.84
Fine-TOM 20540.860.680.56 113150.930.840.78 26960.950.880.81 20180.930.810.72 27380.930.910.90
KGMatcher 19090.890.690.56 107640.940.850.77 25770.980.890.82 14930.940.750.62 23210.940.880.83
LogMap 22550.840.590.46 116480.890.820.76 24910.880.810.75 15770.940.790.68 21880.940.840.75
LSMatch 81.000.000.00 211.000.000.00 241.000.010.00 111.000.010.00 281.000.020.01
OTMapOnto 00.000.000.00 1490.500.000.00 1710.670.010.01 661.000.010.00 1040.760.020.01
TOM 1190.930.080.04 10010.960.150.08 2480.870.160.09 360.900.030.02 2500.940.180.10
Wiktionary 30950.860.760.68 134650.920.910.90 34080.920.920.93 21910.920.830.75 28220.930.920.92

Class results

All matchers were able to generate class correspondences.
marvelcinematicuniverse-marvel memoryalpha-memorybeta memoryalpha-stexpanded starwars-swg starwars-swtor
SizePrec.F-m.Rec. SizePrec.F-m.Rec. SizePrec.F-m.Rec. SizePrec.F-m.Rec. SizePrec.F-m.Rec.
ALOD2Vec 71.001.001.00 211.000.440.29 291.000.760.62 141.000.750.60 291.000.930.87
AMD 00.000.000.00 211.000.440.29 251.000.760.62 00.000.000.00 00.000.000.00
AML 81.001.001.00 360.910.800.71 321.000.820.69 121.000.890.80 301.000.930.87
ATMatcher 111.001.001.00 390.830.770.71 341.000.870.77 131.000.750.60 311.000.930.87
BaselineAltLabel 81.001.001.00 191.000.440.29 191.000.630.46 91.000.570.40 271.000.890.80
BaselineLabel 81.001.001.00 191.000.440.29 191.000.630.46 91.000.570.40 271.000.890.80
Fine-TOM 81.001.001.00 221.000.440.29 271.000.760.62 111.000.750.60 281.000.890.80
KGMatcher 81.001.001.00 271.000.440.29 291.000.700.54 221.000.750.60 301.000.930.87
LogMap 101.001.001.00 210.880.640.50 260.780.640.54 121.000.890.80 281.000.850.73
LSMatch 81.001.001.00 211.000.440.29 241.000.700.54 111.000.750.60 281.000.890.80
OTMapOnto 00.000.000.00 1490.500.530.57 1710.670.710.77 661.001.001.00 1040.760.810.87
TOM 81.001.001.00 211.000.440.29 261.000.760.62 131.000.890.80 291.000.930.87
Wiktionary 81.001.001.00 261.000.440.29 331.000.760.62 141.000.750.60 291.000.930.87

Property results

While last year, most matchers struggled on creating mappings typed as rdf:Property (instead of owl:ObjectProperty or owl:DatatypeProperty, many matchers were able to produce property mappings also in that case.
marvelcinematicuniverse-marvel memoryalpha-memorybeta memoryalpha-stexpanded starwars-swg starwars-swtor
SizePrec.F-m.Rec. SizePrec.F-m.Rec. SizePrec.F-m.Rec. SizePrec.F-m.Rec. SizePrec.F-m.Rec.
ALOD2Vec 211.001.001.00 1110.840.880.92 880.870.910.95 531.001.001.00 1110.980.980.98
AMD 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00
AML 91.000.530.36 840.760.730.70 640.890.850.80 121.000.400.25 730.980.820.71
ATMatcher 240.910.910.91 1030.980.950.92 850.950.950.95 611.001.001.00 1211.000.990.98
BaselineAltLabel 71.000.530.36 411.000.510.34 460.970.800.68 421.001.001.00 1031.000.940.89
BaselineLabel 71.000.530.36 411.000.510.34 460.970.800.68 421.001.001.00 1031.000.940.89
Fine-TOM 00.000.000.00 00.000.000.00 00.000.000.00 421.001.001.00 1031.000.940.89
KGMatcher 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00
LogMap 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00
LSMatch 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00
OTMapOnto 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00
TOM 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00
Wiktionary 221.001.001.00 1120.840.880.92 890.870.910.95 591.001.001.00 1170.980.980.98

Instance results

Only AMD, LSMatch, and OTMapOnto did not return any instance matches.
marvelcinematicuniverse-marvel memoryalpha-memorybeta memoryalpha-stexpanded starwars-swg starwars-swtor
SizePrec.F-m.Rec. SizePrec.F-m.Rec. SizePrec.F-m.Rec. SizePrec.F-m.Rec. SizePrec.F-m.Rec.
ALOD2Vec 30650.860.760.68 133230.920.910.90 32850.920.930.93 21180.920.820.75 26760.920.920.91
AMD 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00
AML 46700.850.680.56 183190.910.890.87 36990.930.930.93 34910.900.810.75 38350.930.920.90
ATMatcher 34810.660.580.52 128640.960.930.91 31620.960.940.92 21690.930.830.76 26190.940.920.91
BaselineAltLabel 25590.860.760.68 134540.880.890.89 31650.880.900.93 16610.920.740.62 25350.920.910.90
BaselineLabel 18640.900.690.56 104920.950.850.77 25170.980.910.84 11940.950.670.52 21420.950.890.84
Fine-TOM 20460.860.680.56 112930.930.850.78 26690.950.890.83 19650.930.810.72 26070.930.910.90
KGMatcher 19010.890.690.57 107370.940.850.78 25480.980.910.84 14710.940.760.63 22910.940.900.86
LogMap 22450.840.600.46 116270.890.820.76 24650.880.820.77 15650.940.800.69 21600.940.860.78
LSMatch 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00
OTMapOnto 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00 00.000.000.00
TOM 1110.930.070.04 9800.960.150.08 2220.860.160.09 230.880.030.01 2210.940.170.10
Wiktionary 30650.860.760.68 133270.920.910.90 32860.920.930.93 21180.920.820.75 26760.920.920.91

Runtime

marvelcinematicuniverse-marvel memoryalpha-memorybeta memoryalpha-stexpanded starwars-swg starwars-swtor
ALOD2Vec 00:05:25 00:03:25 00:02:08 00:05:16 00:05:37
AMD 00:00:00 00:23:39 00:14:08 00:00:00 00:00:00
AML 00:14:43 00:16:34 00:05:24 00:06:54 00:06:49
ATMatcher 00:04:43 00:03:34 00:02:08 00:04:41 00:04:27
BaselineAltLabel 00:02:45 00:01:54 00:01:11 00:02:54 00:02:50
BaselineLabel 00:02:40 00:01:50 00:01:11 00:02:52 00:02:51
Fine-TOM 06:47:31 05:03:51 01:35:05 00:57:10 00:31:31
KGMatcher 00:47:19 00:50:48 00:31:31 01:28:55 01:16:57
LogMap 00:36:24 00:06:09 00:03:50 00:09:24 00:08:57
LSMatch 00:31:04 00:19:00 00:11:53 00:30:34 00:30:22
OTMapOnto 00:00:00 00:09:57 00:06:12 00:16:12 00:16:04
TOM 08:01:39 08:48:33 02:11:48 03:07:06 01:21:18
Wiktionary 00:12:03 00:05:14 00:05:38 00:09:47 00:10:34

Organizers

References

[1] Sven Hertling, Heiko Paulheim: The knowledge graph track at OAEI : Gold standards, baselines, and the golden hammer bias. ESWC 2020. [pdf]

[2] Sven Hertling, Heiko Paulheim: DBkWik: A Consolidated Knowledge Graph from Thousands of Wikis. International Conference on Big Knowledge 2018. [pdf]

[3] Alexandra Hofmann, Samresh Perchani, Jan Portisch, Sven Hertling, and Heiko Paulheim. DBkWik: Towards Knowledge Graph Creation from Thousands of Wikis. International Semantic Web Conference (Posters & Demos) 2017. [pdf]