Evaluating the Performance of Large Language Models for Geometry and Simulation File Generation in Physics-Based Simulations

Research output: Contribution to journalArticlepeer-review

4 Downloads (Pure)

Abstract

Finite-element simulations and computer-aided design workflows require complex preprocessing, with geometry creation and simulation setup traditionally demanding significant manual expertise. The question emerges: can machine learning, namely large language models, help automate these processes? This study evaluates how well nine large language models can automate finite-element simulations starting from natural language prompts, generating both the geometry files for meshing (using Gmsh, an open-source geometry and mesh generator) and the input files needed for the solver (using Elmer, an open-source multiphysics simulation tool). Two standard test cases, a simple bar and a wheel and axle assembly, are used to evaluate and compare their performance. A set of criteria and a scoring system is introduced to assess performance across geometry and simulation setup, covering aspects such as file completeness, Boolean operations, shape fidelity, and displacement error. Results show that most LLMs excel at generating solver input files, achieving 78–88% success rate with
Original languageEnglish
Article number12114
Number of pages15
JournalApplied Sciences
Volume15
Issue number22
Early online date13 Nov 2025
DOIs
Publication statusE-pub ahead of print - 13 Nov 2025

Keywords

  • geometry
  • large language models
  • physics-based simulations
  • LLAMA
  • artificial intelligence in engineering design
  • Gmsh
  • GPT-4

Fingerprint

Dive into the research topics of 'Evaluating the Performance of Large Language Models for Geometry and Simulation File Generation in Physics-Based Simulations'. Together they form a unique fingerprint.

Cite this