volume 415 pages 295-316

On hyperparameter optimization of machine learning algorithms: Theory and practice

Publication typeJournal Article
Publication date2020-11-01
scimago Q1
wos Q1
SJR1.471
CiteScore13.6
Impact factor6.5
ISSN09252312, 18728286
Computer Science Applications
Artificial Intelligence
Cognitive Neuroscience
Abstract
Machine learning algorithms have been used widely in various applications and areas. To fit a machine learning model into different problems, its hyper-parameters must be tuned. Selecting the best hyper-parameter configuration for machine learning models has a direct impact on the model's performance. It often requires deep knowledge of machine learning algorithms and appropriate hyper-parameter optimization techniques. Although several automatic optimization techniques exist, they have different strengths and drawbacks when applied to different types of problems. In this paper, optimizing the hyper-parameters of common machine learning models is studied. We introduce several state-of-the-art optimization techniques and discuss how to apply them to machine learning algorithms. Many available libraries and frameworks developed for hyper-parameter optimization problems are provided, and some open challenges of hyper-parameter optimization research are also discussed in this paper. Moreover, experiments are conducted on benchmark datasets to compare the performance of different optimization methods and provide practical examples of hyper-parameter optimization. This survey paper will help industrial users, data analysts, and researchers to better develop machine learning models by identifying the proper hyper-parameter configurations effectively.
Found 
Found 

Top-30

Journals

10
20
30
40
50
60
70
80
90
IEEE Access
82 publications, 3.48%
Scientific Reports
41 publications, 1.74%
Applied Sciences (Switzerland)
32 publications, 1.36%
Engineering Applications of Artificial Intelligence
29 publications, 1.23%
Lecture Notes in Networks and Systems
26 publications, 1.1%
Sensors
23 publications, 0.98%
Lecture Notes in Computer Science
22 publications, 0.93%
Remote Sensing
21 publications, 0.89%
Communications in Computer and Information Science
20 publications, 0.85%
Expert Systems with Applications
16 publications, 0.68%
Electronics (Switzerland)
15 publications, 0.64%
Applied Soft Computing Journal
15 publications, 0.64%
Water (Switzerland)
14 publications, 0.59%
Materials Today Communications
14 publications, 0.59%
Energies
12 publications, 0.51%
Neural Computing and Applications
12 publications, 0.51%
Neurocomputing
12 publications, 0.51%
PLoS ONE
12 publications, 0.51%
Results in Engineering
12 publications, 0.51%
Mathematics
11 publications, 0.47%
Journal of Cleaner Production
11 publications, 0.47%
Sustainability
10 publications, 0.42%
Knowledge-Based Systems
10 publications, 0.42%
Applied Energy
10 publications, 0.42%
Computers and Chemical Engineering
9 publications, 0.38%
Information (Switzerland)
9 publications, 0.38%
Biomedical Signal Processing and Control
9 publications, 0.38%
Measurement: Journal of the International Measurement Confederation
9 publications, 0.38%
Ocean Engineering
9 publications, 0.38%
10
20
30
40
50
60
70
80
90

Publishers

100
200
300
400
500
600
700
800
Elsevier
723 publications, 30.71%
Springer Nature
451 publications, 19.16%
Institute of Electrical and Electronics Engineers (IEEE)
380 publications, 16.14%
MDPI
277 publications, 11.77%
Wiley
81 publications, 3.44%
Taylor & Francis
51 publications, 2.17%
SAGE
37 publications, 1.57%
Association for Computing Machinery (ACM)
30 publications, 1.27%
American Chemical Society (ACS)
28 publications, 1.19%
Frontiers Media S.A.
26 publications, 1.1%
IOP Publishing
19 publications, 0.81%
American Society of Civil Engineers (ASCE)
15 publications, 0.64%
Public Library of Science (PLoS)
14 publications, 0.59%
Cold Spring Harbor Laboratory
14 publications, 0.59%
AIP Publishing
13 publications, 0.55%
Oxford University Press
13 publications, 0.55%
Hindawi Limited
12 publications, 0.51%
Royal Society of Chemistry (RSC)
10 publications, 0.42%
IGI Global
9 publications, 0.38%
Research Square Platform LLC
9 publications, 0.38%
Emerald
7 publications, 0.3%
Walter de Gruyter
6 publications, 0.25%
PeerJ
5 publications, 0.21%
Institution of Engineering and Technology (IET)
5 publications, 0.21%
JMIR Publications
5 publications, 0.21%
American Geophysical Union
5 publications, 0.21%
EDP Sciences
4 publications, 0.17%
American Physical Society (APS)
4 publications, 0.17%
Society of Petroleum Engineers
4 publications, 0.17%
100
200
300
400
500
600
700
800
  • We do not take into account publications without a DOI.
  • Statistics recalculated weekly.

Are you a researcher?

Create a profile to get free access to personal recommendations for colleagues and new articles.
Metrics
2.4k
Share
Cite this
GOST |
Cite this
GOST Copy
Li Y., Shami A. On hyperparameter optimization of machine learning algorithms: Theory and practice // Neurocomputing. 2020. Vol. 415. pp. 295-316.
GOST all authors (up to 50) Copy
Li Y., Shami A. On hyperparameter optimization of machine learning algorithms: Theory and practice // Neurocomputing. 2020. Vol. 415. pp. 295-316.
RIS |
Cite this
RIS Copy
TY - JOUR
DO - 10.1016/j.neucom.2020.07.061
UR - https://doi.org/10.1016/j.neucom.2020.07.061
TI - On hyperparameter optimization of machine learning algorithms: Theory and practice
T2 - Neurocomputing
AU - Li, Yang
AU - Shami, Abdallah
PY - 2020
DA - 2020/11/01
PB - Elsevier
SP - 295-316
VL - 415
SN - 0925-2312
SN - 1872-8286
ER -
BibTex
Cite this
BibTex (up to 50 authors) Copy
@article{2020_Li,
author = {Yang Li and Abdallah Shami},
title = {On hyperparameter optimization of machine learning algorithms: Theory and practice},
journal = {Neurocomputing},
year = {2020},
volume = {415},
publisher = {Elsevier},
month = {nov},
url = {https://doi.org/10.1016/j.neucom.2020.07.061},
pages = {295--316},
doi = {10.1016/j.neucom.2020.07.061}
}