The goal of the benchmark test is to offer a set of tests which are wide in feature coverage, progressive and stable. It serves the purpose of evaluating the strength and weakness of matchers (by being progressive and wide coverage) and measuring the progress of matchers (by being stable and reusable over the years).
Expected results of matchers are alignments between the named and properties in the namespace of the ontologies. Instances are not in the reference alignments.
The benchmark test library consists of data sets that are built from reference ontologies of different sizes and from different domains.
Since OAEI 2011, Benchmark test suites (or data sets) have been generated from seed ontologies . This year, in addition to the bibliographic ontology, we will generate tests with one or two other ontologies (may be only made available after evaluation):
These two ontologies had a comparable size. The only benefit of the film test is that it was not disclosed to participants before (it actually was chosen after systems were submitted). The following table summarizes the information about ontologies' sizes. classes+prop instances entities triples
It is not necessary anymore to download the data sets (it has always been better to get it on the web). The SEALS platform will provide the data sets.
All these data sets maintain the structure explained in the Example of a complete benchmark data set section, and testing with those data sets can be done by using the SEALS client. This client iterates over tests in a data set whose identifier is provided as a parameter. In all cases, the ontologies found in the data set directories are matched (either against the ontology found in 101/onto.rdf, or both ontologies to match are in the same directory). The resulting alignments must be outputted in the alignment format. They are placed in a local directory given also as a parameter to the client.
The identifiers of data sets for testing with the SEALS client are given below:
We encourage you to use the Alignment API for manipulating and generating your alignments, and, in particular, for computing evaluation of your results.
Contact address is Jerome : Euzenat # inria : fr
 Jérôme Euzenat, Maria Roşoiu, Cássia Trojahn dos Santos. Ontology matching benchmarks: generation, stability, and discriminability, Journal of web semantics 21:30-48, 2013 [DOI:10.1016/j.websem.2013.05.002]