volume 174 pages 106226

Generalization analysis of deep CNNs under maximum correntropy criterion

Publication typeJournal Article
Publication date2024-06-01
scimago Q1
wos Q1
SJR1.491
CiteScore10.6
Impact factor6.3
ISSN08936080, 18792782
Artificial Intelligence
Cognitive Neuroscience
Abstract
Convolutional neural networks (CNNs) have gained immense popularity in recent years, finding their utility in diverse fields such as image recognition, natural language processing, and bio-informatics. Despite the remarkable progress made in deep learning theory, most studies on CNNs, especially in regression tasks, tend to heavily rely on the least squares loss function. However, there are situations where such learning algorithms may not suffice, particularly in the presence of heavy-tailed noises or outliers. This predicament emphasizes the necessity of exploring alternative loss functions that can handle such scenarios more effectively, thereby unleashing the true potential of CNNs. In this paper, we investigate the generalization error of deep CNNs with the rectified linear unit (ReLU) activation function for robust regression problems within an information-theoretic learning framework. Our study demonstrates that when the regression function exhibits an additive ridge structure and the noise possesses a finite pth moment, the empirical risk minimization scheme, generated by the maximum correntropy criterion and deep CNNs, achieves fast convergence rates. Notably, these rates align with the mini-max optimal convergence rates attained by fully connected neural network model with the Huber loss function up to a logarithmic factor. Additionally, we further establish the convergence rates of deep CNNs under the maximum correntropy criterion when the regression function resides in a Sobolev space on the sphere.
Found 
Found 

Top-30

Journals

1
Journal of Complexity
1 publication, 25%
Agriculture (Switzerland)
1 publication, 25%
IEEE Network
1 publication, 25%
Symmetry
1 publication, 25%
1

Publishers

1
2
MDPI
2 publications, 50%
Elsevier
1 publication, 25%
Institute of Electrical and Electronics Engineers (IEEE)
1 publication, 25%
1
2
  • We do not take into account publications without a DOI.
  • Statistics recalculated weekly.

Are you a researcher?

Create a profile to get free access to personal recommendations for colleagues and new articles.
Metrics
4
Share
Cite this
GOST |
Cite this
GOST Copy
Zhang Y. et al. Generalization analysis of deep CNNs under maximum correntropy criterion // Neural Networks. 2024. Vol. 174. p. 106226.
GOST all authors (up to 50) Copy
Zhang Y., Fang Z., Jiang F., Fan J. Generalization analysis of deep CNNs under maximum correntropy criterion // Neural Networks. 2024. Vol. 174. p. 106226.
RIS |
Cite this
RIS Copy
TY - JOUR
DO - 10.1016/j.neunet.2024.106226
UR - https://linkinghub.elsevier.com/retrieve/pii/S0893608024001503
TI - Generalization analysis of deep CNNs under maximum correntropy criterion
T2 - Neural Networks
AU - Zhang, Yingqiao
AU - Fang, Zhiying
AU - Jiang, Fan
AU - Fan, Jun
PY - 2024
DA - 2024/06/01
PB - Elsevier
SP - 106226
VL - 174
PMID - 38490117
SN - 0893-6080
SN - 1879-2782
ER -
BibTex
Cite this
BibTex (up to 50 authors) Copy
@article{2024_Zhang,
author = {Yingqiao Zhang and Zhiying Fang and Fan Jiang and Jun Fan},
title = {Generalization analysis of deep CNNs under maximum correntropy criterion},
journal = {Neural Networks},
year = {2024},
volume = {174},
publisher = {Elsevier},
month = {jun},
url = {https://linkinghub.elsevier.com/retrieve/pii/S0893608024001503},
pages = {106226},
doi = {10.1016/j.neunet.2024.106226}
}