PROMPT ENGINEERING FOR REQUIREMENTS ELICITATION: A COMPARATIVE EVALUATION OF EIGHT TECHNIQUES USING O1

Publikation: Beiträge in SammelwerkenAufsätze in KonferenzbändenForschungbegutachtet

Standard

PROMPT ENGINEERING FOR REQUIREMENTS ELICITATION: A COMPARATIVE EVALUATION OF EIGHT TECHNIQUES USING O1. / Schiller, Kata Amanda; Seibel, Arthur.
51st Design Automation Conference (DAC). Hrsg. / Guang Hao; Dongming Gan; Mark Plecnik; Bingling Huang. The American Society of Mechanical Engineers(ASME), 2025. v03bt03a047 (Proceedings of the ASME Design Engineering Technical Conference; Band 3B-2025).

Publikation: Beiträge in SammelwerkenAufsätze in KonferenzbändenForschungbegutachtet

Harvard

Schiller, KA & Seibel, A 2025, PROMPT ENGINEERING FOR REQUIREMENTS ELICITATION: A COMPARATIVE EVALUATION OF EIGHT TECHNIQUES USING O1. in G Hao, D Gan, M Plecnik & B Huang (Hrsg.), 51st Design Automation Conference (DAC)., v03bt03a047, Proceedings of the ASME Design Engineering Technical Conference, Bd. 3B-2025, The American Society of Mechanical Engineers(ASME), ASME 2025 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, IDETC-CIE 2025, Anaheim, USA / Vereinigte Staaten, 17.08.25. https://doi.org/10.1115/DETC2025-169451

APA

Schiller, K. A., & Seibel, A. (2025). PROMPT ENGINEERING FOR REQUIREMENTS ELICITATION: A COMPARATIVE EVALUATION OF EIGHT TECHNIQUES USING O1. In G. Hao, D. Gan, M. Plecnik, & B. Huang (Hrsg.), 51st Design Automation Conference (DAC) Artikel v03bt03a047 (Proceedings of the ASME Design Engineering Technical Conference; Band 3B-2025). The American Society of Mechanical Engineers(ASME). https://doi.org/10.1115/DETC2025-169451

Vancouver

Schiller KA, Seibel A. PROMPT ENGINEERING FOR REQUIREMENTS ELICITATION: A COMPARATIVE EVALUATION OF EIGHT TECHNIQUES USING O1. in Hao G, Gan D, Plecnik M, Huang B, Hrsg., 51st Design Automation Conference (DAC). The American Society of Mechanical Engineers(ASME). 2025. v03bt03a047. (Proceedings of the ASME Design Engineering Technical Conference). doi: 10.1115/DETC2025-169451

Bibtex

@inbook{0ae2ea51d61647338c23f5198f3592cb,
title = "PROMPT ENGINEERING FOR REQUIREMENTS ELICITATION: A COMPARATIVE EVALUATION OF EIGHT TECHNIQUES USING O1",
abstract = "Large language models (LLMs) offer new opportunities for automating requirements elicitation in early-stage product development. However, many existing approaches rely on complex architectures or fine-tuning, which are resource-intensive and often inaccessible to non-experts. This reveals a critical gap: the demand for lightweight, broadly applicable methods that support effective requirements generation without the need for specialized model training. Prompt engineering addresses this gap by guiding pre-trained LLMs through carefully crafted input instructions, thereby avoiding the need for additional training. This study evaluates eight prompt engineering techniques for generating requirements documents using OpenAI{\textquoteright}s o1 model on the example of a smart electronic product. The analysis focuses on the structure of requirements lists and examines various methods, ranging from basic approaches (zero-shot, one-shot, and few-shot prompting), to reasoning-based techniques (chain-of-thought and tree-of-thought prompting), and advanced strategies (self-consistency, ensemble refinement, and choice shuffle ensemble). The outputs are assessed based on multiple categories using three LLMs, namely Llama-3.2-3B-Instruct, DeepSeek-R1-Distill-Qwen-7BGGUF, and Mistral-7B-Instruct-v0.3. The results show substantial variation in performance. Simpler prompts often produce incomplete or inconsistent outputs, while more advanced strategies show higher quality and coherence. In particular, combining few-shot prompting with tree-of-thought prompting and ensemble refinement produces the most reliable and well-structured requirements lists. These techniques enable the model to reason more transparently and iteratively refine its outputs. However, challenges remain such as susceptibility to hallucinations and a strong dependence on prompt design.",
keywords = "Large Language Models, Product Development, Prompt Engineering, Requirements Elicitation, Engineering",
author = "Schiller, {Kata Amanda} and Arthur Seibel",
note = "Publisher Copyright: Copyright {\textcopyright} 2025 by ASME.; ASME 2025 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, IDETC-CIE 2025 ; Conference date: 17-08-2025 Through 20-08-2025",
year = "2025",
doi = "10.1115/DETC2025-169451",
language = "English",
isbn = "978-0-7918-8923-7 ",
series = "Proceedings of the ASME Design Engineering Technical Conference",
publisher = "The American Society of Mechanical Engineers(ASME)",
editor = "Guang Hao and Dongming Gan and Mark Plecnik and Bingling Huang",
booktitle = "51st Design Automation Conference (DAC)",
address = "United States",

}

RIS

TY - CHAP

T1 - PROMPT ENGINEERING FOR REQUIREMENTS ELICITATION

T2 - ASME 2025 International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, IDETC-CIE 2025

AU - Schiller, Kata Amanda

AU - Seibel, Arthur

N1 - Publisher Copyright: Copyright © 2025 by ASME.

PY - 2025

Y1 - 2025

N2 - Large language models (LLMs) offer new opportunities for automating requirements elicitation in early-stage product development. However, many existing approaches rely on complex architectures or fine-tuning, which are resource-intensive and often inaccessible to non-experts. This reveals a critical gap: the demand for lightweight, broadly applicable methods that support effective requirements generation without the need for specialized model training. Prompt engineering addresses this gap by guiding pre-trained LLMs through carefully crafted input instructions, thereby avoiding the need for additional training. This study evaluates eight prompt engineering techniques for generating requirements documents using OpenAI’s o1 model on the example of a smart electronic product. The analysis focuses on the structure of requirements lists and examines various methods, ranging from basic approaches (zero-shot, one-shot, and few-shot prompting), to reasoning-based techniques (chain-of-thought and tree-of-thought prompting), and advanced strategies (self-consistency, ensemble refinement, and choice shuffle ensemble). The outputs are assessed based on multiple categories using three LLMs, namely Llama-3.2-3B-Instruct, DeepSeek-R1-Distill-Qwen-7BGGUF, and Mistral-7B-Instruct-v0.3. The results show substantial variation in performance. Simpler prompts often produce incomplete or inconsistent outputs, while more advanced strategies show higher quality and coherence. In particular, combining few-shot prompting with tree-of-thought prompting and ensemble refinement produces the most reliable and well-structured requirements lists. These techniques enable the model to reason more transparently and iteratively refine its outputs. However, challenges remain such as susceptibility to hallucinations and a strong dependence on prompt design.

AB - Large language models (LLMs) offer new opportunities for automating requirements elicitation in early-stage product development. However, many existing approaches rely on complex architectures or fine-tuning, which are resource-intensive and often inaccessible to non-experts. This reveals a critical gap: the demand for lightweight, broadly applicable methods that support effective requirements generation without the need for specialized model training. Prompt engineering addresses this gap by guiding pre-trained LLMs through carefully crafted input instructions, thereby avoiding the need for additional training. This study evaluates eight prompt engineering techniques for generating requirements documents using OpenAI’s o1 model on the example of a smart electronic product. The analysis focuses on the structure of requirements lists and examines various methods, ranging from basic approaches (zero-shot, one-shot, and few-shot prompting), to reasoning-based techniques (chain-of-thought and tree-of-thought prompting), and advanced strategies (self-consistency, ensemble refinement, and choice shuffle ensemble). The outputs are assessed based on multiple categories using three LLMs, namely Llama-3.2-3B-Instruct, DeepSeek-R1-Distill-Qwen-7BGGUF, and Mistral-7B-Instruct-v0.3. The results show substantial variation in performance. Simpler prompts often produce incomplete or inconsistent outputs, while more advanced strategies show higher quality and coherence. In particular, combining few-shot prompting with tree-of-thought prompting and ensemble refinement produces the most reliable and well-structured requirements lists. These techniques enable the model to reason more transparently and iteratively refine its outputs. However, challenges remain such as susceptibility to hallucinations and a strong dependence on prompt design.

KW - Large Language Models

KW - Product Development

KW - Prompt Engineering

KW - Requirements Elicitation

KW - Engineering

UR - http://www.scopus.com/inward/record.url?scp=105024317303&partnerID=8YFLogxK

U2 - 10.1115/DETC2025-169451

DO - 10.1115/DETC2025-169451

M3 - Article in conference proceedings

AN - SCOPUS:105024317303

SN - 978-0-7918-8923-7

T3 - Proceedings of the ASME Design Engineering Technical Conference

BT - 51st Design Automation Conference (DAC)

A2 - Hao, Guang

A2 - Gan, Dongming

A2 - Plecnik, Mark

A2 - Huang, Bingling

PB - The American Society of Mechanical Engineers(ASME)

Y2 - 17 August 2025 through 20 August 2025

ER -

DOI