Gerbil – Benchmarking named entity recognition and linking consistently

Publikation: Beiträge in ZeitschriftenZeitschriftenaufsätzeForschungbegutachtet

Standard

Gerbil – Benchmarking named entity recognition and linking consistently. / Röder, Michael; Usbeck, Ricardo; Ngonga Ngomo, Axel Cyrille.
in: Semantic Web, Jahrgang 9, Nr. 5, 2018.

Publikation: Beiträge in ZeitschriftenZeitschriftenaufsätzeForschungbegutachtet

Harvard

APA

Vancouver

Röder M, Usbeck R, Ngonga Ngomo AC. Gerbil – Benchmarking named entity recognition and linking consistently. Semantic Web. 2018;9(5). doi: 10.3233/sw-170286

Bibtex

@article{da05985e99fb43258bf73e7e9d3009e7,
title = "Gerbil – Benchmarking named entity recognition and linking consistently",
abstract = "The ability to compare systems from the same domain is of central importance for their introduction into complex applications. In the domains of named entity recognition and entity linking, the large number of systems and their orthogonal evaluation w.r.t. measures and datasets has led to an unclear landscape regarding the abilities and weaknesses of the different approaches. We present GERBIL—an improved platform for repeatable, storable and citable semantic annotation experiments—and its extension since being release. GERBIL has narrowed this evaluation gap by generating concise, archivable, human- and machine-readable experiments, analytics and diagnostics. The rationale behind our framework is to provide developers, end users and researchers with easy-to-use interfaces that allow for the agile, fine-grained and uniform evaluation of annotation tools on multiple datasets. By these means, we aim to ensure that both tool developers and end users can derive meaningful insights into the extension, integration and use of annotation applications. In particular, GERBIL provides comparable results to tool developers, simplifying the discovery of strengths and weaknesses of their implementations with respect to the state-of-the-art. With the permanent experiment URIs provided by our framework, we ensure the reproducibility and archiving of evaluation results. Moreover, the framework generates data in a machine-processable format, allowing for the efficient querying and post-processing of evaluation results. Additionally, the tool diagnostics provided by GERBIL provide insights into the areas where tools need further refinement, thus allowing developers to create an informed agenda for extensions and end users to detect the right tools for their purposes. Finally, we implemented additional types of experiments including entity typing. GERBIL aims to become a focal point for the state-of-the-art, driving the research agenda of the community by presenting comparable objective evaluation results. Furthermore, we tackle the central problem of the evaluation of entity linking, i.e., we answer the question of how an evaluation algorithm can compare two URIs to each other without being bound to a specific knowledge base. Our approach to this problem opens a way to address the deprecation of URIs of existing gold standards for named entity recognition and entity linking, a feature which is currently not supported by the state-of-the-art. We derived the importance of this feature from usage and dataset requirements collected from the GERBIL user community, which has already carried out more than 24.000 single evaluations using our framework. Through the resulting updates, GERBIL now supports 8 tasks, 46 datasets and 20 systems.",
keywords = "Archivability, Benchmarking framework, Disambiguation, Linking, Named entity recognition, Reusability, Semantic entity annotation system, Informatics, Business informatics",
author = "Michael R{\"o}der and Ricardo Usbeck and {Ngonga Ngomo}, {Axel Cyrille}",
note = "This work was supported by the German Federal Ministry of Education and Research under the project number 03WKCJ4D and the Eurostars projects DIESEL (E!9367) and QAMEL (E!9725) as well as the European Union{\textquoteright}s H2020 research and innovation action HOBBIT under the Grant Agreement number 688227. Publisher Copyright: {\textcopyright} 2018 IOS Press. All rights reserved.",
year = "2018",
doi = "10.3233/sw-170286",
language = "English",
volume = "9",
journal = "Semantic Web",
issn = "1570-0844",
publisher = "SAGE Publications Inc.",
number = "5",

}

RIS

TY - JOUR

T1 - Gerbil – Benchmarking named entity recognition and linking consistently

AU - Röder, Michael

AU - Usbeck, Ricardo

AU - Ngonga Ngomo, Axel Cyrille

N1 - This work was supported by the German Federal Ministry of Education and Research under the project number 03WKCJ4D and the Eurostars projects DIESEL (E!9367) and QAMEL (E!9725) as well as the European Union’s H2020 research and innovation action HOBBIT under the Grant Agreement number 688227. Publisher Copyright: © 2018 IOS Press. All rights reserved.

PY - 2018

Y1 - 2018

N2 - The ability to compare systems from the same domain is of central importance for their introduction into complex applications. In the domains of named entity recognition and entity linking, the large number of systems and their orthogonal evaluation w.r.t. measures and datasets has led to an unclear landscape regarding the abilities and weaknesses of the different approaches. We present GERBIL—an improved platform for repeatable, storable and citable semantic annotation experiments—and its extension since being release. GERBIL has narrowed this evaluation gap by generating concise, archivable, human- and machine-readable experiments, analytics and diagnostics. The rationale behind our framework is to provide developers, end users and researchers with easy-to-use interfaces that allow for the agile, fine-grained and uniform evaluation of annotation tools on multiple datasets. By these means, we aim to ensure that both tool developers and end users can derive meaningful insights into the extension, integration and use of annotation applications. In particular, GERBIL provides comparable results to tool developers, simplifying the discovery of strengths and weaknesses of their implementations with respect to the state-of-the-art. With the permanent experiment URIs provided by our framework, we ensure the reproducibility and archiving of evaluation results. Moreover, the framework generates data in a machine-processable format, allowing for the efficient querying and post-processing of evaluation results. Additionally, the tool diagnostics provided by GERBIL provide insights into the areas where tools need further refinement, thus allowing developers to create an informed agenda for extensions and end users to detect the right tools for their purposes. Finally, we implemented additional types of experiments including entity typing. GERBIL aims to become a focal point for the state-of-the-art, driving the research agenda of the community by presenting comparable objective evaluation results. Furthermore, we tackle the central problem of the evaluation of entity linking, i.e., we answer the question of how an evaluation algorithm can compare two URIs to each other without being bound to a specific knowledge base. Our approach to this problem opens a way to address the deprecation of URIs of existing gold standards for named entity recognition and entity linking, a feature which is currently not supported by the state-of-the-art. We derived the importance of this feature from usage and dataset requirements collected from the GERBIL user community, which has already carried out more than 24.000 single evaluations using our framework. Through the resulting updates, GERBIL now supports 8 tasks, 46 datasets and 20 systems.

AB - The ability to compare systems from the same domain is of central importance for their introduction into complex applications. In the domains of named entity recognition and entity linking, the large number of systems and their orthogonal evaluation w.r.t. measures and datasets has led to an unclear landscape regarding the abilities and weaknesses of the different approaches. We present GERBIL—an improved platform for repeatable, storable and citable semantic annotation experiments—and its extension since being release. GERBIL has narrowed this evaluation gap by generating concise, archivable, human- and machine-readable experiments, analytics and diagnostics. The rationale behind our framework is to provide developers, end users and researchers with easy-to-use interfaces that allow for the agile, fine-grained and uniform evaluation of annotation tools on multiple datasets. By these means, we aim to ensure that both tool developers and end users can derive meaningful insights into the extension, integration and use of annotation applications. In particular, GERBIL provides comparable results to tool developers, simplifying the discovery of strengths and weaknesses of their implementations with respect to the state-of-the-art. With the permanent experiment URIs provided by our framework, we ensure the reproducibility and archiving of evaluation results. Moreover, the framework generates data in a machine-processable format, allowing for the efficient querying and post-processing of evaluation results. Additionally, the tool diagnostics provided by GERBIL provide insights into the areas where tools need further refinement, thus allowing developers to create an informed agenda for extensions and end users to detect the right tools for their purposes. Finally, we implemented additional types of experiments including entity typing. GERBIL aims to become a focal point for the state-of-the-art, driving the research agenda of the community by presenting comparable objective evaluation results. Furthermore, we tackle the central problem of the evaluation of entity linking, i.e., we answer the question of how an evaluation algorithm can compare two URIs to each other without being bound to a specific knowledge base. Our approach to this problem opens a way to address the deprecation of URIs of existing gold standards for named entity recognition and entity linking, a feature which is currently not supported by the state-of-the-art. We derived the importance of this feature from usage and dataset requirements collected from the GERBIL user community, which has already carried out more than 24.000 single evaluations using our framework. Through the resulting updates, GERBIL now supports 8 tasks, 46 datasets and 20 systems.

KW - Archivability

KW - Benchmarking framework

KW - Disambiguation

KW - Linking

KW - Named entity recognition

KW - Reusability

KW - Semantic entity annotation system

KW - Informatics

KW - Business informatics

UR - http://www.scopus.com/inward/record.url?scp=85066125676&partnerID=8YFLogxK

U2 - 10.3233/sw-170286

DO - 10.3233/sw-170286

M3 - Journal articles

AN - SCOPUS:85066125676

VL - 9

JO - Semantic Web

JF - Semantic Web

SN - 1570-0844

IS - 5

ER -

DOI

Zuletzt angesehen

Publikationen

  1. Chapter 9: Particular Remedies for Non-performance: Section 2: Withholding Performance
  2. Informatik
  3. Making an impression with open strategy
  4. Wavlet analysis for rotor fault diagnosis with rub-impact of rotor systems
  5. The relationship between values and knowledge in visioning for landscape management
  6. Differentiating Different Types of Cognitive Load
  7. Explorations in social spaces
  8. Performance of methods to select landscape metrics for modelling species richness
  9. “Circuits of Commons”: Exploring the Connections Between Economic Lives and the Commons
  10. Learning from partially annotated sequences
  11. Implementing UNESCO's Convention on Cultural Diversity at the regional level
  12. Internet-based public debate of CCS
  13. Introduction
  14. Efficient co-regularised least squares regression
  15. Influence of data clouds fusion from 3D real-time vision system on robotic group dead reckoning in unknown terrain
  16. Exchanging Knowledge and Good Practices of Education for Sustainable Development within a Global Student Organization (oikos)
  17. Quantification and analysis of surface macroplastic contamination on arable areas
  18. Modeling Grounding Processes in Chat-based CSCL
  19. Searching for New Languages, Searching for Minor Voices in the Archive
  20. Forest structure and heterogeneity increase diversity and alter composition of host–parasitoid networks
  21. Influence of Mg content in Al alloys on processing characteristics and dynamically recrystallized microstructure of friction surfacing deposits
  22. Introduction to the Special Issue Section
  23. Safer Spaces
  24. RelHunter
  25. Using measures of reading time regularity (RTR) to quantify eye movement dynamics, and how they are shaped by linguistic information
  26. Exploding Images
  27. Embedding Evidence on Conservation Interventions Within a Context of Multilevel Governance
  28. Discussion report part 2
  29. Priority Rule-based Planning Approaches for Regeneration Processes
  30. Action rate models for predicting actions in soccer