Generalization analysis of deep CNNs under maximum correntropy criterion
Publication type: Journal Article
Publication date: 2024-06-01
scimago Q1
wos Q1
SJR: 1.491
CiteScore: 10.6
Impact factor: 6.3
ISSN: 08936080, 18792782
PubMed ID:
38490117
Artificial Intelligence
Cognitive Neuroscience
Abstract
Convolutional neural networks (CNNs) have gained immense popularity in recent years, finding their utility in diverse fields such as image recognition, natural language processing, and bio-informatics. Despite the remarkable progress made in deep learning theory, most studies on CNNs, especially in regression tasks, tend to heavily rely on the least squares loss function. However, there are situations where such learning algorithms may not suffice, particularly in the presence of heavy-tailed noises or outliers. This predicament emphasizes the necessity of exploring alternative loss functions that can handle such scenarios more effectively, thereby unleashing the true potential of CNNs. In this paper, we investigate the generalization error of deep CNNs with the rectified linear unit (ReLU) activation function for robust regression problems within an information-theoretic learning framework. Our study demonstrates that when the regression function exhibits an additive ridge structure and the noise possesses a finite pth moment, the empirical risk minimization scheme, generated by the maximum correntropy criterion and deep CNNs, achieves fast convergence rates. Notably, these rates align with the mini-max optimal convergence rates attained by fully connected neural network model with the Huber loss function up to a logarithmic factor. Additionally, we further establish the convergence rates of deep CNNs under the maximum correntropy criterion when the regression function resides in a Sobolev space on the sphere.
Found
Nothing found, try to update filter.
Found
Nothing found, try to update filter.
Top-30
Journals
|
1
|
|
|
Journal of Complexity
1 publication, 25%
|
|
|
Agriculture (Switzerland)
1 publication, 25%
|
|
|
IEEE Network
1 publication, 25%
|
|
|
Symmetry
1 publication, 25%
|
|
|
1
|
Publishers
|
1
2
|
|
|
MDPI
2 publications, 50%
|
|
|
Elsevier
1 publication, 25%
|
|
|
Institute of Electrical and Electronics Engineers (IEEE)
1 publication, 25%
|
|
|
1
2
|
- We do not take into account publications without a DOI.
- Statistics recalculated weekly.
Are you a researcher?
Create a profile to get free access to personal recommendations for colleagues and new articles.
Metrics
4
Total citations:
4
Citations from 2024:
4
(100%)
Cite this
GOST |
RIS |
BibTex
Cite this
GOST
Copy
Zhang Y. et al. Generalization analysis of deep CNNs under maximum correntropy criterion // Neural Networks. 2024. Vol. 174. p. 106226.
GOST all authors (up to 50)
Copy
Zhang Y., Fang Z., Jiang F., Fan J. Generalization analysis of deep CNNs under maximum correntropy criterion // Neural Networks. 2024. Vol. 174. p. 106226.
Cite this
RIS
Copy
TY - JOUR
DO - 10.1016/j.neunet.2024.106226
UR - https://linkinghub.elsevier.com/retrieve/pii/S0893608024001503
TI - Generalization analysis of deep CNNs under maximum correntropy criterion
T2 - Neural Networks
AU - Zhang, Yingqiao
AU - Fang, Zhiying
AU - Jiang, Fan
AU - Fan, Jun
PY - 2024
DA - 2024/06/01
PB - Elsevier
SP - 106226
VL - 174
PMID - 38490117
SN - 0893-6080
SN - 1879-2782
ER -
Cite this
BibTex (up to 50 authors)
Copy
@article{2024_Zhang,
author = {Yingqiao Zhang and Zhiying Fang and Fan Jiang and Jun Fan},
title = {Generalization analysis of deep CNNs under maximum correntropy criterion},
journal = {Neural Networks},
year = {2024},
volume = {174},
publisher = {Elsevier},
month = {jun},
url = {https://linkinghub.elsevier.com/retrieve/pii/S0893608024001503},
pages = {106226},
doi = {10.1016/j.neunet.2024.106226}
}