Gerbil – Benchmarking named entity recognition and linking consistently

Research output: Journal contributionsJournal articlesResearchpeer-review

Authors

The ability to compare systems from the same domain is of central importance for their introduction into complex applications. In the domains of named entity recognition and entity linking, the large number of systems and their orthogonal evaluation w.r.t. measures and datasets has led to an unclear landscape regarding the abilities and weaknesses of the different approaches. We present GERBIL—an improved platform for repeatable, storable and citable semantic annotation experiments—and its extension since being release. GERBIL has narrowed this evaluation gap by generating concise, archivable, human- and machine-readable experiments, analytics and diagnostics. The rationale behind our framework is to provide developers, end users and researchers with easy-to-use interfaces that allow for the agile, fine-grained and uniform evaluation of annotation tools on multiple datasets. By these means, we aim to ensure that both tool developers and end users can derive meaningful insights into the extension, integration and use of annotation applications. In particular, GERBIL provides comparable results to tool developers, simplifying the discovery of strengths and weaknesses of their implementations with respect to the state-of-the-art. With the permanent experiment URIs provided by our framework, we ensure the reproducibility and archiving of evaluation results. Moreover, the framework generates data in a machine-processable format, allowing for the efficient querying and post-processing of evaluation results. Additionally, the tool diagnostics provided by GERBIL provide insights into the areas where tools need further refinement, thus allowing developers to create an informed agenda for extensions and end users to detect the right tools for their purposes. Finally, we implemented additional types of experiments including entity typing. GERBIL aims to become a focal point for the state-of-the-art, driving the research agenda of the community by presenting comparable objective evaluation results. Furthermore, we tackle the central problem of the evaluation of entity linking, i.e., we answer the question of how an evaluation algorithm can compare two URIs to each other without being bound to a specific knowledge base. Our approach to this problem opens a way to address the deprecation of URIs of existing gold standards for named entity recognition and entity linking, a feature which is currently not supported by the state-of-the-art. We derived the importance of this feature from usage and dataset requirements collected from the GERBIL user community, which has already carried out more than 24.000 single evaluations using our framework. Through the resulting updates, GERBIL now supports 8 tasks, 46 datasets and 20 systems.

Original languageEnglish
JournalSemantic Web
Volume9
Issue number5
ISSN1570-0844
DOIs
Publication statusPublished - 2018
Externally publishedYes

Bibliographical note

This work was supported by the German Federal Ministry of Education and Research under the project number 03WKCJ4D and the Eurostars projects DIESEL (E!9367) and QAMEL (E!9725) as well as the European Union’s H2020 research and innovation action HOBBIT under the Grant Agreement number 688227.

Publisher Copyright:
© 2018 IOS Press. All rights reserved.

    Research areas

  • Archivability, Benchmarking framework, Disambiguation, Linking, Named entity recognition, Reusability, Semantic entity annotation system
  • Informatics
  • Business informatics

DOI

Recently viewed

Publications

  1. Children's use of spatial skills in solving two map-reading tasks in real space.
  2. The temporal pattern of creativity and implementation in teams
  3. Legitimizing Digital Transformation: From System Integration to Platformization
  4. Assessing authenticity in modelling test items: deriving a theoretical model
  5. Modeling self-determination theory motivation data by using unfolding IRT
  6. The Influence of Robots’ Emotion Expressions on the Uncanny-Valley-Effect
  7. Migration-Based Multilingualism in the English as a Foreign Language Classroom
  8. A Playful Approach to Interactive Media in the Foreign Language Classroom
  9. Study of Single Filament Dielectric Barrier Discharge in Argon
  10. Framework for empirical research on science teaching and learning
  11. Artificial Intelligence in Foreign Language Learning and Teaching
  12. A Multilevel CFA–MTMM Approach for Multisource Feedback Instruments
  13. Microsimulation - A survey of principles, developments and applications
  14. Depression-specific Costs and their Factors based on SHI Routine data
  15. Performance of the DSM-5-based criteria for Internet addiction
  16. A comprehensive Eulerian modeling framework for airborne mercury species
  17. Quasi-in-situ observation of microstructure at the friction interface
  18. Short run comovement, persistent shocks and the business cycle
  19. Explaining Investment Dynamics: Empirical Evidence from German New Ventures
  20. Editorial: Machine Learning and Data Mining in Materials Science
  21. Optimization of a gaseous multitube detector for soft X-ray detection
  22. The Economic Value of Clickstream Data From an Advertiser's Perspective
  23. Path dependence and technological expectations in transport policy
  24. Data practices in apps from Brazil: What do privacy policies inform us about?
  25. Utilization of protein-rich residues in biotechnological processes
  26. Swarm Robotics, or: the smartness of 'a bunch of cheap dumb things'
  27. Chapter 9: Particular Remedies for Non-performance: Section 4: Price Reduction
  28. User Authentication via Multifaceted Mouse Movements and Outlier Exposure
  29. A Method to Enhance the Accuracy of Time of Flight Measurement Systems
  30. Detection of significant tracer gases by means of polymer gas sensors
  31. From simulation to real-world robotic mobile fulfillment systems
  32. Developing shaping competence in informal setting at universities