Szczegóły publikacji
Opis bibliograficzny
Open and closed source models for LLM-generated metaheuristics solving engineering optimization problem / Roman Senkerik, Adam Viktorin, Tomas Kadavy, Jozef Kovac, Peter Janku, Libor Pekar, Hubert GUZOWSKI, Maciej SMOŁKA, Aleksander BYRSKI, Michal PLUHÁČEK // W: Applications of evolutionary computation : 28th European conference, EvoApplications 2025 : held as part of EvoStar 2025 : Trieste, Italy, April 23–25, 2025 : proceedings, Pt. 2 / eds. Pablo García-Sánchez, Emma Hart, Sarah L. Thomson. — Cham : Springer Nature Switzerland, cop. 2025. — (Lecture Notes in Computer Science ; ISSN 0302-9743 ; vol. 15613). — ISBN: 978-3-031-90064-8; e-ISBN: 978-3-031-90065-5. — S. 372–385. — Bibliogr., Abstr. — Publikacja dostępna online od: 2025-04-17
Autorzy (10)
- Senkerik Roman
- Viktorin Adam
- Kadavy Tomas
- Kovac Jozef
- Janku Peter
- Pekar Libor
- AGHGuzowski Hubert
- AGHSmołka Maciej
- AGHByrski Aleksander
- AGHPluháček Michal
Dane bibliometryczne
| ID BaDAP | 159982 |
|---|---|
| Data dodania do BaDAP | 2025-06-24 |
| DOI | 10.1007/978-3-031-90065-5_23 |
| Rok publikacji | 2025 |
| Typ publikacji | materiały konferencyjne (aut.) |
| Otwarty dostęp | |
| Wydawca | Springer |
| Konferencja | International Conference on Applications of Evolutionary Computation 2025 |
| Czasopismo/seria | Lecture Notes in Computer Science |
Abstract
This paper explores the applicability of generative AI (genAI), specifically Large Language Models (LLMs), for the automatic generation and configuration of metaheuristic algorithms to address a real-world engineering problem: the optimal parameter estimation of time-delay systems in interconnected heating-cooling loops. The study introduces a pioneering workflow and iterative architecture with feedback within the emerging field of genAI-driven optimization for real optimization problems, eliminating the need for manually crafted or modified algorithms. This automated system empowers domain experts in engineering to solve complex optimization problems with minimal knowledge of optimization algorithms, lowering the barrier to entry for sophisticated algorithm use. We demonstrate how LLMs can generate effective optimizers under conditions like connstrained optimization problems where the solution lies near the boundaries of the search space. Four state-of-the-art LLMs (closed and open-sourced) have been selected for experiments. These are GPT-4o, GPT-4o mini, Claude Sonnet 3.5 and Llama 3.1. All studied LLMs generated metaheuristics that outperformed the initialization baseline optimization method (Random Search and CMA-ES). Notably, the Claude Sonnet 3.5 model generated a metaheuristic with the best mean results, almost matching the performance of the tuned state-of-the-art DISH algorithm, as an example of adaptive Differential Evolution.