Reporting and Analysing the Environmental Impact of Language Models on the Example of Commonsense Question Answering with External Knowledge

Research output: Contributions to collected editions/worksArticle in conference proceedingsResearch

Standard

Reporting and Analysing the Environmental Impact of Language Models on the Example of Commonsense Question Answering with External Knowledge. / Usmanova, Aida; Huang, Junbo; Banerjee, Debayan et al.
Sustainable AI Conference 2023: Sustainable AI Across Borders: Conference Proceedings. Vol. abs/2408.01453 2024.

Research output: Contributions to collected editions/worksArticle in conference proceedingsResearch

Harvard

Usmanova, A, Huang, J, Banerjee, D & Usbeck, R 2024, Reporting and Analysing the Environmental Impact of Language Models on the Example of Commonsense Question Answering with External Knowledge. in Sustainable AI Conference 2023: Sustainable AI Across Borders: Conference Proceedings. vol. abs/2408.01453, 2. Sustainable AI Conference 2023, Bonn, North Rhine-Westphalia, Germany, 30.05.23. https://doi.org/10.48550/ARXIV.2408.01453

APA

Usmanova, A., Huang, J., Banerjee, D., & Usbeck, R. (2024). Reporting and Analysing the Environmental Impact of Language Models on the Example of Commonsense Question Answering with External Knowledge. Manuscript in preparation. In Sustainable AI Conference 2023: Sustainable AI Across Borders: Conference Proceedings (Vol. abs/2408.01453) https://doi.org/10.48550/ARXIV.2408.01453

Vancouver

Usmanova A, Huang J, Banerjee D, Usbeck R. Reporting and Analysing the Environmental Impact of Language Models on the Example of Commonsense Question Answering with External Knowledge. In Sustainable AI Conference 2023: Sustainable AI Across Borders: Conference Proceedings. Vol. abs/2408.01453. 2024 doi: 10.48550/ARXIV.2408.01453

Bibtex

@inbook{4f4ea465d3704eceba687547285bf745,
title = "Reporting and Analysing the Environmental Impact of Language Models on the Example of Commonsense Question Answering with External Knowledge",
abstract = "Human-produced emissions are growing at an alarming rate, causing already observable changes in the climate and environment in general. Each year global carbon dioxide emissions hit a new record, and it is reported that 0.5% of total US greenhouse gas emissions are attributed to data centres as of 2021. The release of ChatGPT in late 2022 sparked social interest in Large Language Models (LLMs), the new generation of Language Models with a large number of parameters and trained on massive amounts of data. Currently, numerous companies are releasing products featuring various LLMs, with many more models in development and awaiting release. Deep Learning research is a competitive field, with only models that reach top performance attracting attention and being utilized. Hence, achieving better accuracy and results is often the first priority, while the model's efficiency and the environmental impact of the study are neglected. However, LLMs demand substantial computational resources and are very costly to train, both financially and environmentally. It becomes essential to raise awareness and promote conscious decisions about algorithmic and hardware choices. Providing information on training time, the approximate carbon dioxide emissions and power consumption would assist future studies in making necessary adjustments and determining the compatibility of available computational resources with model requirements. In this study, we infused T5 LLM with external knowledge and fine-tuned the model for Question-Answering task. Furthermore, we calculated and reported the approximate environmental impact for both steps. The findings demonstrate that the smaller models may not always be sustainable options, and increased training does not always imply better performance. The most optimal outcome is achieved by carefully considering both performance and efficiency factors.",
keywords = "Informatics",
author = "Aida Usmanova and Junbo Huang and Debayan Banerjee and Ricardo Usbeck",
year = "2024",
doi = "10.48550/ARXIV.2408.01453",
language = "English",
volume = "abs/2408.01453",
booktitle = "Sustainable AI Conference 2023: Sustainable AI Across Borders",
note = "2. Sustainable AI Conference 2023 : Sustainable AI Across Borders ; Conference date: 30-05-2023 Through 01-06-2023",
url = "https://www.uni-bonn.de/de/veranstaltungen/sustainable-ai-conference-2023-sustainable-ai-across-borders",

}

RIS

TY - CHAP

T1 - Reporting and Analysing the Environmental Impact of Language Models on the Example of Commonsense Question Answering with External Knowledge

AU - Usmanova, Aida

AU - Huang, Junbo

AU - Banerjee, Debayan

AU - Usbeck, Ricardo

N1 - Conference code: 2

PY - 2024

Y1 - 2024

N2 - Human-produced emissions are growing at an alarming rate, causing already observable changes in the climate and environment in general. Each year global carbon dioxide emissions hit a new record, and it is reported that 0.5% of total US greenhouse gas emissions are attributed to data centres as of 2021. The release of ChatGPT in late 2022 sparked social interest in Large Language Models (LLMs), the new generation of Language Models with a large number of parameters and trained on massive amounts of data. Currently, numerous companies are releasing products featuring various LLMs, with many more models in development and awaiting release. Deep Learning research is a competitive field, with only models that reach top performance attracting attention and being utilized. Hence, achieving better accuracy and results is often the first priority, while the model's efficiency and the environmental impact of the study are neglected. However, LLMs demand substantial computational resources and are very costly to train, both financially and environmentally. It becomes essential to raise awareness and promote conscious decisions about algorithmic and hardware choices. Providing information on training time, the approximate carbon dioxide emissions and power consumption would assist future studies in making necessary adjustments and determining the compatibility of available computational resources with model requirements. In this study, we infused T5 LLM with external knowledge and fine-tuned the model for Question-Answering task. Furthermore, we calculated and reported the approximate environmental impact for both steps. The findings demonstrate that the smaller models may not always be sustainable options, and increased training does not always imply better performance. The most optimal outcome is achieved by carefully considering both performance and efficiency factors.

AB - Human-produced emissions are growing at an alarming rate, causing already observable changes in the climate and environment in general. Each year global carbon dioxide emissions hit a new record, and it is reported that 0.5% of total US greenhouse gas emissions are attributed to data centres as of 2021. The release of ChatGPT in late 2022 sparked social interest in Large Language Models (LLMs), the new generation of Language Models with a large number of parameters and trained on massive amounts of data. Currently, numerous companies are releasing products featuring various LLMs, with many more models in development and awaiting release. Deep Learning research is a competitive field, with only models that reach top performance attracting attention and being utilized. Hence, achieving better accuracy and results is often the first priority, while the model's efficiency and the environmental impact of the study are neglected. However, LLMs demand substantial computational resources and are very costly to train, both financially and environmentally. It becomes essential to raise awareness and promote conscious decisions about algorithmic and hardware choices. Providing information on training time, the approximate carbon dioxide emissions and power consumption would assist future studies in making necessary adjustments and determining the compatibility of available computational resources with model requirements. In this study, we infused T5 LLM with external knowledge and fine-tuned the model for Question-Answering task. Furthermore, we calculated and reported the approximate environmental impact for both steps. The findings demonstrate that the smaller models may not always be sustainable options, and increased training does not always imply better performance. The most optimal outcome is achieved by carefully considering both performance and efficiency factors.

KW - Informatics

UR - https://dblp.org/db/journals/corr/index.html

UR - https://www.sustainable-ai-conference.eu/

U2 - 10.48550/ARXIV.2408.01453

DO - 10.48550/ARXIV.2408.01453

M3 - Article in conference proceedings

VL - abs/2408.01453

BT - Sustainable AI Conference 2023: Sustainable AI Across Borders

T2 - 2. Sustainable AI Conference 2023

Y2 - 30 May 2023 through 1 June 2023

ER -

Recently viewed

Researchers

  1. Marcus Erben

Publications

  1. Measurement invariance in a grid-based measure of academic self-concept
  2. A direct test of the similarity assumption — Focusing on differences as compared with similarities decreases automatic imitation
  3. Empowered or informed? Seeking to mitigate gender differences in first-offer assertiveness through pre-negotiation interventions
  4. How to support students-learning in mathematical bridging-courses using ITS? Remedial Scenarios in the EU-Project Math-Bridge
  5. Priority effects of time of arrival of plant functional groups override sowing interval or density effects
  6. Beyond Structural Adjustment
  7. Identification of hybridization strategies for combining fused filament fabrication with unidirectional tape reinforcement
  8. Insights into adoption of farming practices through multiple lenses
  9. A Robust Approximated Derivative Action of a PID Regulator to be Applied in a Permanent Magnet Synchronous Motor Control
  10. Next level production networks
  11. Problem-Based and Project-Based Learning for Sustainable Development
  12. Multifractality Versus (Mono-) Fractality as Evidence of Nonlinear Interactions Across Timescales
  13. Improving mechanical properties of chip-based aluminum extrudates by integrated extrusion and equal channel angular pressing (iECAP)
  14. Multiple
  15. Instruments for research on transition. Applied methods and approaches for exploring the transition of young care leavers to adulthood
  16. Complex Trait-Treatment-Interaction analysis
  17. PI Control Applied to a Small-Scale Thermal System with Heating and Cooling Sources
  18. An Exploration of humans‘ ability to recognize emotions displayed by robots
  19. Sustainable development and learning for sustainability through a regional network project
  20. Integration of laboratory experiments into introductory electrical engineering courses
  21. Active plasma resonance spectroscopy: Eigenfunction solutions in spherical geometry
  22. Machine Learning and Data Mining for Sports Analytics
  23. Question answering over linked data
  24. Analog, Digital, and the Cybernetic Illusion
  25. Assessing mire-specific biodiversity with an indicator based approach
  26. Influence of three different unstable shoe constructions on EMG-activity during treadmill walking