Fine-Tuned LLMs vs. Rule-Based NLP for UML Diagram Generation: An Educational Evaluation
Author(s) -
Reshma P Nair,
M.G. Thushara,
Vijayan Sugumaran
Publication year - 2025
Publication title -
ieee access
Language(s) - English
Resource type - Magazines
SCImago Journal Rank - 0.587
H-Index - 127
eISSN - 2169-3536
DOI - 10.1109/access.2025.3638372
Subject(s) - aerospace , bioengineering , communication, networking and broadcast technologies , components, circuits, devices and systems , computing and processing , engineered materials, dielectrics and plasmas , engineering profession , fields, waves and electromagnetics , general topics for engineers , geoscience , nuclear engineering , photonics and electrooptics , power, energy and industry applications , robotics and control systems , signal processing and analysis , transportation
The automatic generation of Unified Modeling Language (UML) diagrams from natural language requirements continues to be a difficult task in software engineering education and practice. Conventional rule-based Natural Language Processing (NLP) approaches often fail to handle semantic ambiguity and typically demand extensive manual adjustments. Recent developments in Large Language Models (LLMs) demonstrate stronger contextual reasoning abilities, but their capability to produce accurate and structured diagrams-particularly in instructional settings—has not been thoroughly examined. This work compares a traditional NLP pipeline with a fine-tuned Qwen2.5-Coder-7B-Instruct model using a dataset of 75 annotated specifications. The fine-tuned model attained a semantic fidelity of 94%, a G-Eval correctness score of 0.78, and reduced manual correction requirements by 78%. In contrast, the NLP pipeline achieved 67% fidelity and required a 52% correction effort. Statistical analysis using MANOVA and ANOVA confirmed that these differences across the student, NLP, and LLM groups were statistically significant (p < 0.001). Moreover, students working with LLM-generated diagrams reported a 30% decrease in effort and greater engagement during modeling tasks. The findings demonstrate that fine-tuned LLMs can substantially improve the precision and interpretability of automated UML diagram generation while supporting more interactive and effective learning in software engineering contexts.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom