Skip to main content

2024 | OriginalPaper | Buchkapitel

LPSD: Low-Rank Plus Sparse Decomposition for Highly Compressed CNN Models

verfasst von : Kuei-Hsiang Huang, Cheng-Yu Sie, Jhong-En Lin, Che-Rung Lee

Erschienen in: Advances in Knowledge Discovery and Data Mining

Verlag: Springer Nature Singapore

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Low-rank decomposition that explores and eliminates the linear dependency within a tensor is often used as a structured model pruning method for deep convolutional neural networks. However, the model accuracy declines rapidly as the compression ratio increases over a threshold. We have observed that with a small amount of sparse elements, the model accuracy can be recovered significantly for the highly compressed CNN models. Based on this premise, we developed a novel method, called LPSD (Low-rank Plus Sparse Decomposition), that decomposes a CNN weight tensor into a combination of a low-rank and a sparse components, which can better maintain the accuracy for the high compression ratio. For a pretrained model, the network structure of each layer is split into two branches: one for low-rank part and one for sparse part. LPSD adapts the alternating approximation algorithm to minimize the global error and the local error alternatively. An exhausted search method with pruning is designed to search the optimal group number, ranks, and sparsity. Experimental results demonstrate that in most scenarios, LPSD achieves better accuracy compared to the state-of-the-art methods when the model is highly compressed.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Anhänge
Nur mit Berechtigung zugänglich
Literatur
1.
Zurück zum Zitat Cai, J.F., Li, J., Xia, D.: Generalized low-rank plus sparse tensor estimation by fast Riemannian optimization (2022) Cai, J.F., Li, J., Xia, D.: Generalized low-rank plus sparse tensor estimation by fast Riemannian optimization (2022)
2.
Zurück zum Zitat Chu, B.S., Lee, C.R.: Low-rank tensor decomposition for compression of convolutional neural networks using funnel regularization (2021) Chu, B.S., Lee, C.R.: Low-rank tensor decomposition for compression of convolutional neural networks using funnel regularization (2021)
4.
Zurück zum Zitat Han, S., et al.: DSD: Dense-sparse-dense training for deep neural networks (2017) Han, S., et al.: DSD: Dense-sparse-dense training for deep neural networks (2017)
5.
Zurück zum Zitat Hawkins, C., Yang, H., Li, M., Lai, L., Chandra, V.: Low-rank+sparse tensor compression for neural networks (2021) Hawkins, C., Yang, H., Li, M., Lai, L., Chandra, V.: Low-rank+sparse tensor compression for neural networks (2021)
6.
Zurück zum Zitat Huang, W., et al.: Deep low-rank plus sparse network for dynamic MR imaging (2021) Huang, W., et al.: Deep low-rank plus sparse network for dynamic MR imaging (2021)
8.
Zurück zum Zitat Kaloshin, P.: Convolutional neural networks compression with low rank and sparse tensor decompositions (2020) Kaloshin, P.: Convolutional neural networks compression with low rank and sparse tensor decompositions (2020)
9.
Zurück zum Zitat Kim, Y.D., Park, E., Yoo, S., Choi, T., Yang, L., Shin, D.: Compression of deep convolutional neural networks for fast and low power mobile applications (2016) Kim, Y.D., Park, E., Yoo, S., Choi, T., Yang, L., Shin, D.: Compression of deep convolutional neural networks for fast and low power mobile applications (2016)
10.
Zurück zum Zitat Liang, C.C., Lee, C.R.: Automatic selection of tensor decomposition for compressing convolutional neural networks a case study on VGG-type networks. In: 2021 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW), pp. 770–778 (2021). https://doi.org/10.1109/IPDPSW52791.2021.00115 Liang, C.C., Lee, C.R.: Automatic selection of tensor decomposition for compressing convolutional neural networks a case study on VGG-type networks. In: 2021 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW), pp. 770–778 (2021). https://​doi.​org/​10.​1109/​IPDPSW52791.​2021.​00115
12.
Zurück zum Zitat Lin, T., Stich, S.U., Barba, L., Dmitriev, D., Jaggi, M.: Dynamic model pruning with feedback (2020) Lin, T., Stich, S.U., Barba, L., Dmitriev, D., Jaggi, M.: Dynamic model pruning with feedback (2020)
13.
Zurück zum Zitat Otazo, R., Candès, E., Sodickson, D.: Low-rank plus sparse matrix decomposition for accelerated dynamic MRI with separation of background and dynamic components. Magn. Reson. Med. 73, 1125–1136 (2014). https://doi.org/10.1002/mrm.25240 Otazo, R., Candès, E., Sodickson, D.: Low-rank plus sparse matrix decomposition for accelerated dynamic MRI with separation of background and dynamic components. Magn. Reson. Med. 73, 1125–1136 (2014). https://​doi.​org/​10.​1002/​mrm.​25240
16.
Zurück zum Zitat Zhang, X., Wang, L., Gu, Q.: A unified framework for low-rank plus sparse matrix recovery (2018) Zhang, X., Wang, L., Gu, Q.: A unified framework for low-rank plus sparse matrix recovery (2018)
Metadaten
Titel
LPSD: Low-Rank Plus Sparse Decomposition for Highly Compressed CNN Models
verfasst von
Kuei-Hsiang Huang
Cheng-Yu Sie
Jhong-En Lin
Che-Rung Lee
Copyright-Jahr
2024
Verlag
Springer Nature Singapore
DOI
https://doi.org/10.1007/978-981-97-2242-6_28

Premium Partner