Szczegóły publikacji
Opis bibliograficzny
Efficient pruning and compression techniques for convolutional neural networks to preserve knowledge and optimize performance / Jakub SKRZYŃSKI, Adrian HORZYK // W: Neural Information Processing : 31st International Conference, ICONIP 2024 : Auckland, New Zealand, December 2–6, 2024: proceedings , Pt. 1 / eds. Mufti Mahmud, [et al.]. — Singapore : Springer Nature Singapore, cop. 2025. — ( Lecture Notes in Computer Science ; ISSN 0302-9743 ; LNCS 15286 ). — ISBN: 978-981-96-6575-4; e-ISBN: 978-981-96-6576-1. — S. 381–395. — Bibliogr., Abstr. — Błąd w nazwisku J. Skrzyński
Autorzy (2)
Dane bibliometryczne
| ID BaDAP | 160772 |
|---|---|
| Data dodania do BaDAP | 2025-08-01 |
| DOI | 10.1007/978-981-96-6576-1_26 |
| Rok publikacji | 2025 |
| Typ publikacji | materiały konferencyjne (aut.) |
| Otwarty dostęp | |
| Wydawca | Springer |
| Konferencja | International Conference on Neural Information Processing 2024 |
| Czasopismo/seria | Lecture Notes in Computer Science |
Abstract
Rapid growth in the complexity and size of Convolutional Neural Networks (CNNs) poses significant challenges in computational resources and energy consumption. This paper presents a novel approach to CNN optimization through a combined pruning and compression technique designed to preserve knowledge and enhance efficiency. Unlike traditional pruning methods that introduce sparsity and require specialized hardware, our method focuses on pruning entire filters based on their importance, followed by an innovative compression strategy that merges the least informative filters. This preserves the knowledge contained within these filters while maintaining the network’s structural integrity. We propose a three-step algorithm: selecting low-information filters using entropy metrics, grouping similar filters, and merging these groups to retain critical information. Our approach significantly reduces the network size without compromising accuracy, as demonstrated using the CIFAR-10, MNIST, Fashion MNIST, and USPS datasets, as well as a modified and original VGG16 architecture. The experiments carried out have shown that our method achieves a substantial reduction in the number of parameters and Floating Point Operations (FLOPs), lowering computational costs by up to 86.82% while preserving up to 99% of the original accuracy of the model. This paper contributes to the field of deep learning by offering a scalable, hardware-agnostic solution for CNN optimization, making it highly suitable for deployment in resource-constrained environments. Future work will explore the application of this method to various architectures and datasets to further validate its efficacy and versatility.