Language Model Transformers as Evaluators for Open-domain Dialogues

Publikation: Beiträge in SammelwerkenAufsätze in KonferenzbändenForschungbegutachtet

Standard

Language Model Transformers as Evaluators for Open-domain Dialogues. / Nedelchev, Rostislav; Lehmann, Jens; Usbeck, Ricardo.
COLING 2020 - 28th International Conference on Computational Linguistics: Proceedings of the Conference. Hrsg. / Donia Scott; Nuria Bel; Chengqing Zong. Association for Computational Linguistics (ACL), 2020. S. 6797-6808 (COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference).

Publikation: Beiträge in SammelwerkenAufsätze in KonferenzbändenForschungbegutachtet

Harvard

Nedelchev, R, Lehmann, J & Usbeck, R 2020, Language Model Transformers as Evaluators for Open-domain Dialogues. in D Scott, N Bel & C Zong (Hrsg.), COLING 2020 - 28th International Conference on Computational Linguistics: Proceedings of the Conference. COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference, Association for Computational Linguistics (ACL), S. 6797-6808, 28th International Conference on Computational Linguistics, COLING 2020, Virtual, Online, Spanien, 08.12.20. https://doi.org/10.18653/v1/2020.coling-main.599

APA

Nedelchev, R., Lehmann, J., & Usbeck, R. (2020). Language Model Transformers as Evaluators for Open-domain Dialogues. In D. Scott, N. Bel, & C. Zong (Hrsg.), COLING 2020 - 28th International Conference on Computational Linguistics: Proceedings of the Conference (S. 6797-6808). (COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.coling-main.599

Vancouver

Nedelchev R, Lehmann J, Usbeck R. Language Model Transformers as Evaluators for Open-domain Dialogues. in Scott D, Bel N, Zong C, Hrsg., COLING 2020 - 28th International Conference on Computational Linguistics: Proceedings of the Conference. Association for Computational Linguistics (ACL). 2020. S. 6797-6808. (COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference). doi: 10.18653/v1/2020.coling-main.599

Bibtex

@inbook{060baa868fe74263b7f5495df8027644,
title = "Language Model Transformers as Evaluators for Open-domain Dialogues",
abstract = "Computer-based systems for communication with humans are a cornerstone of AI research since the 1950s. So far, the most effective way to assess the quality of the dialogues produced by these systems is to use resource-intensive manual labor instead of automated means. In this work, we investigate whether language models (LM) based on transformer neural networks can indicate the quality of a conversation. In a general sense, language models are methods that learn to predict one or more words based on an already given context. Due to their unsupervised nature, they are candidates for efficient, automatic indication of dialogue quality. We demonstrate that human evaluators have a positive correlation between the output of the language models and scores. We also provide some insights into their behavior and inner-working in a conversational context.",
keywords = "Informatics, Business informatics",
author = "Rostislav Nedelchev and Jens Lehmann and Ricardo Usbeck",
note = "We acknowledge the support of the EU projects Cleopatra (GA 812997) and TAILOR (GA 952215), the Federal Ministry for Economic Affairs and Energy (BMWi) project SPEAKER (FKZ 01MK20011A), the German Federal Ministry of Education and Research (BMBF) projects and excellence clusters ML2R (FKZ 01 15 18038 A/B/C), MLwin (01S18050 D/F), ScaDS.AI (01/S18026A) as well as the Fraunhofer Zukunftsstiftung project JOSEPH. Publisher Copyright: {\textcopyright} 2020 COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference. All rights reserved.; 28th International Conference on Computational Linguistics, COLING 2020 ; Conference date: 08-12-2020 Through 13-12-2020",
year = "2020",
month = jan,
day = "1",
doi = "10.18653/v1/2020.coling-main.599",
language = "English",
series = "COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference",
publisher = "Association for Computational Linguistics (ACL)",
pages = "6797--6808",
editor = "Donia Scott and Nuria Bel and Chengqing Zong",
booktitle = "COLING 2020 - 28th International Conference on Computational Linguistics",
address = "United States",
url = "https://coling2020.org, https://coling2020.org/COLING2020programme.pdf",

}

RIS

TY - CHAP

T1 - Language Model Transformers as Evaluators for Open-domain Dialogues

AU - Nedelchev, Rostislav

AU - Lehmann, Jens

AU - Usbeck, Ricardo

N1 - We acknowledge the support of the EU projects Cleopatra (GA 812997) and TAILOR (GA 952215), the Federal Ministry for Economic Affairs and Energy (BMWi) project SPEAKER (FKZ 01MK20011A), the German Federal Ministry of Education and Research (BMBF) projects and excellence clusters ML2R (FKZ 01 15 18038 A/B/C), MLwin (01S18050 D/F), ScaDS.AI (01/S18026A) as well as the Fraunhofer Zukunftsstiftung project JOSEPH. Publisher Copyright: © 2020 COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference. All rights reserved.

PY - 2020/1/1

Y1 - 2020/1/1

N2 - Computer-based systems for communication with humans are a cornerstone of AI research since the 1950s. So far, the most effective way to assess the quality of the dialogues produced by these systems is to use resource-intensive manual labor instead of automated means. In this work, we investigate whether language models (LM) based on transformer neural networks can indicate the quality of a conversation. In a general sense, language models are methods that learn to predict one or more words based on an already given context. Due to their unsupervised nature, they are candidates for efficient, automatic indication of dialogue quality. We demonstrate that human evaluators have a positive correlation between the output of the language models and scores. We also provide some insights into their behavior and inner-working in a conversational context.

AB - Computer-based systems for communication with humans are a cornerstone of AI research since the 1950s. So far, the most effective way to assess the quality of the dialogues produced by these systems is to use resource-intensive manual labor instead of automated means. In this work, we investigate whether language models (LM) based on transformer neural networks can indicate the quality of a conversation. In a general sense, language models are methods that learn to predict one or more words based on an already given context. Due to their unsupervised nature, they are candidates for efficient, automatic indication of dialogue quality. We demonstrate that human evaluators have a positive correlation between the output of the language models and scores. We also provide some insights into their behavior and inner-working in a conversational context.

KW - Informatics

KW - Business informatics

UR - http://www.scopus.com/inward/record.url?scp=85108285068&partnerID=8YFLogxK

UR - https://www.mendeley.com/catalogue/0f9694bb-370d-3c37-bb25-8347d9aac64a/

U2 - 10.18653/v1/2020.coling-main.599

DO - 10.18653/v1/2020.coling-main.599

M3 - Article in conference proceedings

AN - SCOPUS:85108285068

T3 - COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference

SP - 6797

EP - 6808

BT - COLING 2020 - 28th International Conference on Computational Linguistics

A2 - Scott, Donia

A2 - Bel, Nuria

A2 - Zong, Chengqing

PB - Association for Computational Linguistics (ACL)

T2 - 28th International Conference on Computational Linguistics, COLING 2020

Y2 - 8 December 2020 through 13 December 2020

ER -

DOI

Zuletzt angesehen

Publikationen

  1. Old-Age Security and Silver Workers
  2. Book review of Kang-Kwong Luke/Theodossia-Souala Pavlidou: Telephone Calls. Unity and Diversity in Conversational Structure across Languages and Cultures.
  3. Alcohol Affects Goal Commitment by Explicitly and Implicitly Induced Myopia
  4. Christian Steinbacher
  5. Introduction: Children's Literature Global and Local
  6. Migration in der Adoleszenz
  7. Pesticide peak concentration reduction in a small vegetated treatment system controlled by chemograph shape
  8. Logistik und Migration
  9. The Quality of the KombiFiD-Sample of Business Services Enterprises
  10. The human resource practices of small businesses
  11. School leaders’ innovation-related self-efficacy: professional development and learning networks make a difference
  12. Interaction effects of effort-reward imbalance and overcommitment on emotional exhaustion and job performance
  13. Sustainable engineering education in research and practice
  14. Treatment of comorbid alcohol use disorders and depression with cognitive-behavioural therapy and motivational interviewing
  15. Die (Un)Berechenbarkeit des Krieges
  16. Pollinator shortage and global crop yield
  17. Zielablösezyklus
  18. „Nicht naiv genug, an den Fortschritt zu glauben“
  19. Fremdfinanzierung junger Unternehmen
  20. Nietzsche und der Mythos
  21. Einleitung
  22. Digital Health Literacy and Web-Based Information-Seeking Behaviors of University Students in Germany during the COVID-19 Pandemic
  23. “Re-connecting people and nature”: wrong term, wrong goal?
  24. School Will Never End
  25. Teams und ihre Entwicklung
  26. Introduction: Higher Education for Sustainability