Neural Network Compression Based on Tensor Ring Decomposition
2
Department of Electrical and Computer Engineering, The State University of New York at Stony Brook, Stony Brook, NY, USA
|
Publication type: Journal Article
Publication date: 2025-03-01
scimago Q1
wos Q1
SJR: 3.686
CiteScore: 24.7
Impact factor: 8.9
ISSN: 2162237X, 21622388
PubMed ID:
38687669
Abstract
Deep neural networks (DNNs) have made great breakthroughs and seen applications in many domains. However, the incomparable accuracy of DNNs is achieved with the cost of considerable memory consumption and high computational complexity, which restricts their deployment on conventional desktops and portable devices. To address this issue, low-rank factorization, which decomposes the neural network parameters into smaller sized matrices or tensors, has emerged as a promising technique for network compression. In this article, we propose leveraging the emerging tensor ring (TR) factorization to compress the neural network. We investigate the impact of both parameter tensor reshaping and TR decomposition (TRD) on the total number of compressed parameters. To achieve the maximal parameter compression, we propose an algorithm based on prime factorization that simultaneously identifies the optimal tensor reshaping and TRD. In addition, we discover that different execution orders of the core tensors result in varying computational complexities. To identify the optimal execution order, we construct a novel tree structure. Based on this structure, we propose a top-to-bottom splitting algorithm to schedule the execution of core tensors, thereby minimizing computational complexity. We have performed extensive experiments using three kinds of neural networks with three different datasets. The experimental results demonstrate that, compared with the three state-of-the-art algorithms for low-rank factorization, our algorithm can achieve better performance with much lower memory consumption and lower computational complexity.
Found
Nothing found, try to update filter.
Found
Nothing found, try to update filter.
Top-30
Journals
|
1
|
|
|
IEEE Network
1 publication, 12.5%
|
|
|
IEEE Communications Letters
1 publication, 12.5%
|
|
|
IEEE TRANSACTIONS ON SUSTAINABLE COMPUTING
1 publication, 12.5%
|
|
|
Science China Information Sciences
1 publication, 12.5%
|
|
|
IEEE Transactions on Neural Networks and Learning Systems
1 publication, 12.5%
|
|
|
Journal of Imaging
1 publication, 12.5%
|
|
|
Neurocomputing
1 publication, 12.5%
|
|
|
1
|
Publishers
|
1
2
3
4
5
|
|
|
Institute of Electrical and Electronics Engineers (IEEE)
5 publications, 62.5%
|
|
|
Springer Nature
1 publication, 12.5%
|
|
|
MDPI
1 publication, 12.5%
|
|
|
Elsevier
1 publication, 12.5%
|
|
|
1
2
3
4
5
|
- We do not take into account publications without a DOI.
- Statistics recalculated weekly.
Are you a researcher?
Create a profile to get free access to personal recommendations for colleagues and new articles.
Metrics
8
Total citations:
8
Citations from 2024:
7
(87.5%)
Cite this
GOST |
RIS |
BibTex |
MLA
Cite this
GOST
Copy
Xie K. et al. Neural Network Compression Based on Tensor Ring Decomposition // IEEE Transactions on Neural Networks and Learning Systems. 2025. Vol. 36. No. 3. pp. 5388-5402.
GOST all authors (up to 50)
Copy
Xie K., Liu C., Wang X., Li X., Xie G., Wen J., Li K. Neural Network Compression Based on Tensor Ring Decomposition // IEEE Transactions on Neural Networks and Learning Systems. 2025. Vol. 36. No. 3. pp. 5388-5402.
Cite this
RIS
Copy
TY - JOUR
DO - 10.1109/tnnls.2024.3383392
UR - https://ieeexplore.ieee.org/document/10510501/
TI - Neural Network Compression Based on Tensor Ring Decomposition
T2 - IEEE Transactions on Neural Networks and Learning Systems
AU - Xie, Kun
AU - Liu, Can
AU - Wang, Xin
AU - Li, Xiaocan
AU - Xie, Gaogang
AU - Wen, Jigang
AU - Li, Kenli
PY - 2025
DA - 2025/03/01
PB - Institute of Electrical and Electronics Engineers (IEEE)
SP - 5388-5402
IS - 3
VL - 36
PMID - 38687669
SN - 2162-237X
SN - 2162-2388
ER -
Cite this
BibTex (up to 50 authors)
Copy
@article{2025_Xie,
author = {Kun Xie and Can Liu and Xin Wang and Xiaocan Li and Gaogang Xie and Jigang Wen and Kenli Li},
title = {Neural Network Compression Based on Tensor Ring Decomposition},
journal = {IEEE Transactions on Neural Networks and Learning Systems},
year = {2025},
volume = {36},
publisher = {Institute of Electrical and Electronics Engineers (IEEE)},
month = {mar},
url = {https://ieeexplore.ieee.org/document/10510501/},
number = {3},
pages = {5388--5402},
doi = {10.1109/tnnls.2024.3383392}
}
Cite this
MLA
Copy
Xie, Kun, et al. “Neural Network Compression Based on Tensor Ring Decomposition.” IEEE Transactions on Neural Networks and Learning Systems, vol. 36, no. 3, Mar. 2025, pp. 5388-5402. https://ieeexplore.ieee.org/document/10510501/.