UML Sequence Diagram Generation: A Multi-Model, Multi-Domain Evaluation
Journal article, 2025

The automation of Unified Modeling Language (UML) sequence diagram generation has posed a persistent challenge in software engineering, with existing approaches relying heavily on manual processes. Recent advancements in natural language processing (NLP), particularly through large language models (LLMs), offer promising solutions for automating this task. This paper investigates the use of LLMs in automating the generation of UML sequence diagrams from natural language requirements. We evaluate three state-of-the-art LLMs, GPT-4o, Mixtral 8x7B, and Llama 3.1 8B, across multiple datasets, including both public and proprietary requirements, to assess their performance in terms of correctness, completeness, clarity, and readability. The results indicate GPT-4o consistently outperforms the other models in most metrics. Our findings highlight the potential of LLMs to streamline requirements engineering by reducing manual effort, although further refinement is needed to enhance their performance in complex scenarios. This study provides key insights into the strengths and limitations of these models, and offers practical guidance for their application, advancing the understanding of how LLMs can support automation in software engineering tasks.

UML Sequence Diagram

Large Language Models (LLMs)

Requirements Engineering

Artificial Intelligence

Author

Chi Xiao

Ericsson

D. Stahl

Linköping University

Jan Bosch

Chalmers, Computer Science and Engineering (Chalmers), Interaction Design and Software Engineering

University of Gothenburg

IEEE ACM International Conference on Software Engineering Software Engineering in Practice

28327640 (ISSN) 28327659 (eISSN)

2025 272-283

Subject Categories (SSIF 2025)

Software Engineering

DOI

10.1109/ICSE-SEIP66354.2025.00030

More information

Latest update

10/3/2025