Detailed instructions about system wrapping and submission are given below.
The OAEI 2024 campaign will once again confront ontology matchers to ontology and data sources to be matched.
This year, the following test sets are available:
T-Box/Schema matching
- anatomy
- The anatomy
real world case is about matching the Adult Mouse Anatomy (2744 classes) and the NCI Thesaurus (3304 classes) describing the human anatomy.
- conference
-
The goal of the track is to find alignments within a collection of ontologies describing the domain of organising conferences. Additionally, 'complex correspondences' are also very welcome. Alignments will be evaluated automatically against reference alignments also considering its uncertain version presented at ISWC 2014. Summary results along with detail performance results for each ontology pair (test case) and comparison with tools' performance from last years will be provided.
- Multifarm
-
This dataset is composed of a subset of the Conference dataset, translated in nine different languages (Arabic, Chinese, Czech, Dutch, French, German, Portuguese, Russian, and Spanish) and the corresponding alignments between these ontologies. Based on these test cases, it is possible to evaluate and compare the performance of matching approaches with a special focus on multilingualism.
- Complex
-
This track evaluates the detection of complex correspondences between ontologies of the conference domain.
- Food Nutritional Composition
-
This track consists of finding alignments between food concepts from CIQUAL, the French food nutritional composition database, and food concepts from SIREN. Food concepts from both databases are described in LanguaL, a well-known multilingual thesaurus using faceted classification.
- Interactive matching evaluation
(interactive)
-
This track offers the possibility to compare different interactive matching tools which require user interaction.
The goal is to show if user interaction can improve the matching results, which methods are most promising and how many
interactions are necessary. All participating systems are evaluated using an oracle which bases on the reference alignment.
Using the SEALS client, the matching system only needs to be slightly adapted to participate to this track.
- Bio-ML
-
The Bio-ML track is a Machine Learning (ML) friendly Biomedical track for Equivalence and Subsumption Matching.
This track presents an unified evaluation framework suitable for both ML-based and non-ML-based OM systems.
The datasets of this track are based on Mondo and UMLS Metathesarus. This track
supersedes the previous largebio and
phenotype tracks.
- Biodiversity and Ecology (biodiv)
-
The goal of the track is to find pairwise alignments between ontologies and thesauri that are particularly useful for biodiversity and ecology research and are being used in various projects. They have been developed in parallel and are very overlapping. They are semantically rich and contain tens of thousands of classes.
- Digital Humanities (dh)
- The goal of the Digital Humanities track is to evaluate matching system performance when dealing with small datasets in different languages and specialist terms from archaeology, cultural history and the interlink of DH and computer science.
The track offers manually compiled gold standard reference alignments for all the test cases, ensuring semantic, lexical and part-of-speech similartiy.
- Archaeology multiling (arch-multiling)
- The Archaeology multilingual track aims to evaluate if matching systems are capable of finding alingments in archaeological monolingual datasets in different languages and specialist terms.
The track is based on a test case of the Digital Humanities track.
- Circular Economy (CE)
-
The Circular Economy track is about matching relevant Circular Economy ontologies.
Instance and schema matching
- Knowledge graph
-
The Knowledge Graph Track contains nine isolated knowledge graphs with instance and schema data.
The goal of the task is to match both the instances and the schema.
Instance matching (link discovery)
- SPIMBENCH (spimbench)
-
The goal of this track is to determine when two OWL instances describe the same Creative Work.
The datasets are generated and transformed using SPIMBENCH by altering a set of original data through value-based, structure-based, and semantics-aware transformations (simple combination of transformations).
- Link Discovery (link)
- This track proposes a benchmark generator to deal with link discovery for spatial data where spatial data are
represented as trajectories (i.e., sequences of longitude, latitude pairs).
- Pharmacogenomics (pgx)
- The Pharmacogenomics Track involves n-ary tuples representing so-called "pharmacogenomic relationships" and their components of three distinct types: drugs, genetic factors, and phenotypes..
Tabular data to Knowledge Graph matching
- TD→KG (special track)
-
Tabular data to Knowledge Graph (KG) matching is the process of assigning semantic tags from Knowledge Graphs
(e.g., Wikidata or DBpedia) to the elements of a table (e.g., a web table or an arbitrary csv file).
Ontology alignment and link discovery systems are welcome to participate.
We plan to create input data in OWL/RDF format to facilitate their participation.
There will be prizes sponsored by SIRIUS and IBM Research.
Evaluation
Preparation phase
All public datasets should be available by the end of this phase.
MELT includes several built-in evaluation tracks: built-in datasets.
New organisers: can perform the evaluation via a Local Track or contacting the platform chairs to create a new built-in track.
Execution phase
OAEI participants should follow the MELT instructions (schema matching tracks) and/or the HOBBIT instructions (spimbench and link discovery tracks)
depending on the tracks they are willing to participate. The TD→KG track is an exception as it has its own evaluation process.
We encourage systems developers to test their systems in the early stages of this phase to avoid last minute
problems with the evaluation infrastructure. Once the execution phase ends, there will be limited time to solve technical problems
with the evaluation platforms.
Evaluation phase
Evaluation will be run under MELT
or HOBBIT infrastructure.
The TD→KG track is an exception as it has its own evaluation process.
Participants will be evaluated with respect to
all of the OAEI tracks (when possible) even though the system
might be specialized for some specific kind of matching problems.
We know that this can be a problem for some systems that have specifically been developed for, e.g., matching biomedical ontologies;
but this point can still be emphasized in the specific results paper about the system in case the results generated for some specific
track are not good at all.
The results will be reported at the International Workshop on Ontology Matching,
which will be collocated with the 22nd International Semantic Web Conference (ISWC 2024).
OAEI rules
Please note that, a matcher may want to behave differently given what it is
provided with as ontologies; however, this should not be based on features
specific of the tracks (e.g., there is a specific string in the URL, or a specific
class name) but on features of the ontologies (e.g., there are no instances or
labels are in German). Check the OAEI rules here.
Systems that rely or are derived from other ontology matching systems should:
(a) clearly state the system they rely on, and (b) what was changed from / added to the original system.
Withdrawal of systems is possible up to one week after submission. After this period you accept that
your systems will be evaluated and the results will be made publicly available within the OAEI pages and the
OAEI evaluation report in accordance to the
OAEI data policy.
June 30th- preliminary datasets available.
July 31st- preparation phase ends and final datasets are available.
July 31st - participants register their tool (mandatory). Please use this form (requires a google account and a valid email)
August 31st- execution phase ends and participants submit final versions of their tools. MELT tracks (zip or tar.gz file, e.g., LogMap.zip) using this form. HOBBIT tracks (via platform).
September 22nd- alignments submission. This applies only when your tool requires substantial hardware or software resources. Please use this form for submission.
September 30th- evaluation phase ends and results are available.
October 14thOctober 21th- Preliminary version of system papers due. Submit PDF paper (e.g., LogMap_prelim.pdf). Please use this form (requires a google account and a valid email).
- November 11th or 12th
- The 19th Ontology matching workshop
- November 13-15th
- The 23rd International Semantic Web Conference.
- November 20th
- Final version of system papers due. Submit a PDF (e.g., LogMap_final.pdf) paper. Please use this form (requires a google account and a valid email).
From the results of the experiments, participants are expected
to provide the organisers with a paper to be published in the proceedings
of the Ontology Matching workshop.
The paper should be no more than 8 pages long and formatted using the
CEUR Latex template or the CEUR Word template. Long-running systems can submit a 2 pages summary
if there were not significant additions to the system. Please use this form for the submission (requires a google account and a valid email)
These papers are not peer-reviewed, but they will revised by 1-2 OAEI organisers. The main objective of these OAEI paper
is to keep track of the participants and the description of matchers which took part in the campaign.
To ensure easy comparability among the participants it is recommended that the paper follows this
outline:
- Presentation of the system
- State, purpose, general statement
- Specific techniques used
- Adaptations made for the evaluation
- Link to the system and parameters file
- Link to the set of provided alignments (in align format)
- Results
- 2.x) a comment for each dataset performed
- General comments
(not necessaryly by putting the section below but preferably in
this order).
- Comments on the results (strength and weaknesses)
- Discussions on the way to improve the proposed system
- Comments on the OAEI procedure (e.g., comments on the MELT evaluation, if relevant)
- Comments on the OAEI test cases
- Comments on the OAEI measures
- Proposed new measures
- Conclusions
- References