Szczegóły publikacji
Opis bibliograficzny
Constrained hybrid metaheuristic algorithm for probabilistic neural networks learning / Piotr A. KOWALSKI, Szymon KUCHARCZYK, Jacek MAŃDZIUK // Information Sciences ; ISSN 0020-0255. — 2025 — vol. 713 art. no. 122185, s. 1–17. — Bibliogr. s. 16–17, Abstr. — Publikacja dostępna online od: 2025-04-09. — P. A. Kowalski - dod. afiliacja: Systems Research Institute, Polish Academy of Sciences, Warsaw, Poland. — J. Mańdziuk - dod. afiliacja: Faculty of Mathematics and Information Science, Warsaw University of Technology, Warsaw, Poland
Autorzy (3)
Słowa kluczowe
Dane bibliometryczne
| ID BaDAP | 159844 |
|---|---|
| Data dodania do BaDAP | 2025-06-16 |
| Tekst źródłowy | URL |
| DOI | 10.1016/j.ins.2025.122185 |
| Rok publikacji | 2025 |
| Typ publikacji | artykuł w czasopiśmie |
| Otwarty dostęp | |
| Czasopismo/seria | Information Sciences |
Abstract
This study investigates the potential of hybrid metaheuristic algorithms to enhance the training of Probabilistic Neural Networks (PNNs) by leveraging the complementary strengths of multiple optimisation strategies. Traditional learning methods, such as gradient-based approaches, often struggle to optimise high-dimensional and uncertain environments, while single-method metaheuristics may fail to exploit the solution space fully. To address these challenges, we propose the constrained Hybrid Metaheuristic (cHM) algorithm, a novel approach that combines multiple population-based optimisation techniques into a unified framework. The proposed procedure operates in two phases: an initial probing phase evaluates multiple metaheuristics to identify the best-performing one based on the error rate, followed by a fitting phase where the selected metaheuristic refines the PNN to achieve optimal smoothing parameters. This iterative process ensures efficient exploration and convergence, enhancing the network's generalisation and classification accuracy. cHM integrates several popular metaheuristics, such as BAT, Simulated Annealing, Flower Pollination Algorithm, Bacterial Foraging Optimisation, and Particle Swarm Optimisation as internal optimisers. To evaluate cHM performance, experiments were conducted on 16 datasets with varying characteristics, including binary and multiclass classification tasks, balanced and imbalanced class distributions, and diverse feature dimensions. In the experimental evaluation, CHM outperforms individual metaheuristics by a large margin in all the metrics tested. Specifically, the average test accuracy metric is ranked first in 10 out of 16 datasets, with the runner-up method scoring only 3 wins. Similarly, in the average test precision metric, cHM is the winning method for 9 datasets, with the second method securing 3 wins. Finally, in the average test recall metric, the respective figures are 8 wins for CHM and 5 for the second-best method. At the same time, due to its composite nature, CHM's average training fit time is about 2-3 times greater than the average time for the competing single metaheuristics. Overall, the results demonstrate that cHM effectively combines the strengths of individual metaheuristics, leading to faster convergence and more robust learning. The proposed method improves classification performance across diverse datasets by optimising the smoothing parameters of PNNs, proving its application flexibility and efficiency.