An Automatic Benchmark Generator for Entity Recognition and Linking
3 months ago by René Speck
The creation of gold standards is of central importance for the objective assessment and development of approaches in the domain of computer science. The manual creation of gold standards for named entity recognition and entity linking is time and resource intensive. Moreover, recent works show that such gold standards contain a large proportion of mistakes and are difficult to maintain.
At the 11th International Conference on Natural Language Generation (INLG 2018), we presented BENGAL - a novel automatic generation of such gold standards, as a complement to manually created benchmarks. The main advantage of our benchmarks is that they can be readily generated at any time. They are also cost-effective while being guaranteed to be free of annotation errors. We compare the performance of 11 tools on benchmarks in English generated by BENGAL, and on 16 benchmarks created manually. We show that our approach can be ported easily across languages by presenting results achieved by 4 tools on both Brazilian Portuguese and Spanish.
The results suggest that our approach can generate diverse benchmarks with characteristics similar to those of a large proportion of existing benchmarks in several languages. Overall, our results suggest that BENGAL benchmarks can ease the development of named entity recognition and entity linking tools (especially for resource-poor languages) by providing developers with insights into their performance at virtually no cost. Hence, BENGAL can improve the push towards better named entity recognition and entity linking frameworks.