The goal of the benchmark test is to offer a set of tests which are wide in feature coverage, progressive and stable. It serves the purpose of evaluating the strength and weakness of matchers (by being progressive and wide coverage) and measuring the progress of matchers (by being stable and reusable over the years).
Expected results of matchers are alignments between the named and properties in the namespace of the ontologies. Instances are not in the reference alignments.
The benchmark test library consists of data sets that are built from reference ontologies of different sizes and from different domains.
Since OAEI 2011, Benchmark test suites (or data sets) have been generated from seed ontologies . This year, in addition to the bibliographic ontology, we have used one large ontology:
The following table summarizes the information about ontologies' sizes. classes+prop instances entities triples
It is not necessary anymore to download the data sets (it has always been better to get it on the web). The SEALS platform will provide the data sets.
All those data sets maintain the structure explained in the Example of a complete benchmark data set section, and testing with those data sets can be done by using the SEALS client. This client iterates over tests in a data set whose identifier is provided as a parameter. In all cases, the ontologies found in the data set directories are matched (either against the ontology found in 101/onto.rdf, or both ontologies to match are in the same directory). The resulting alignments must be outputted in the alignment format. They are placed in a local directory given also as a parameter to the client.
The identifiers of data sets for testing with the SEALS client are given below:
This last test was replaced by:
We encourage you to use the Alignment API for manipulating and generating your alignments, and, in particular, for computing evaluation of your results.
Contact address is Jerome : Euzenat # inria : fr
 Jérôme Euzenat, Maria Roşoiu, Cássia Trojahn dos Santos. Ontology matching benchmarks: generation, stability, and discriminability, Journal of web semantics 21:30-48, 2013 [DOI:10.1016/j.websem.2013.05.002]