Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization
Optimizing hyper parameters in Convolutional Neural networks is a tedious process for many researchers and practitioners. It requires a high degree of expertise or experience to optimise the hyper parameters, and manual optimisation is likely to be biased. To date, methods or approaches to automate...
Saved in:
Main Authors: | , , |
---|---|
Format: | Book Section |
Language: | English |
Published: |
Computing & Intelligent Systems, SCRS
2022
|
Subjects: | |
Online Access: | http://utpedia.utp.edu.my/id/eprint/24082/1/Comparative%20Study%20of%20Surrogate%20Techniques%20for%20CNN%20Hyperparameter%20Optimization.pdf https://doi.org/10.52458/978-81-95502-00-4-48 http://utpedia.utp.edu.my/id/eprint/24082/ https://www.publications.scrs.in/chapter/978-81-95502-00-4/48 https://doi.org/10.52458/978-81-95502-00-4-48 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
id |
oai:utpedia.utp.edu.my:24082 |
---|---|
record_format |
eprints |
spelling |
oai:utpedia.utp.edu.my:240822023-05-15T07:44:34Z http://utpedia.utp.edu.my/id/eprint/24082/ Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization Mohd Aszemi, Nurshazlyn M. Zakaria, Nordin Paneer Selvam, Dhanapal Durai Dominic T Technology (General) Optimizing hyper parameters in Convolutional Neural networks is a tedious process for many researchers and practitioners. It requires a high degree of expertise or experience to optimise the hyper parameters, and manual optimisation is likely to be biased. To date, methods or approaches to automate hyper parameter optimization include grid search, random search, and Genetic Algorithms (GAs). However, evaluating large number of sample points in the hyperparameter configuration space, as is typically required by these methods, is computationally expensive process. Hence, the objective of this paper is to explore regression as a surrogate technique in CNN hyperparameter optimisation. Performance in terms of accuracy, error rate, training time and coefficient of determination (R2) are evaluated and recorded. Although there is no significant performance difference between the resulting optimized Deep Learning and state-of-the-art on CIFAR-10 datasets, using regression as a surrogate technique for CNN hyperparameter optimization contributes to minimising the time taken for the optimization process, a benefit which has not been fully explored in the literature to the best of the author’s knowledge. Computing & Intelligent Systems, SCRS 2022-05-06 Book Section PeerReviewed text en http://utpedia.utp.edu.my/id/eprint/24082/1/Comparative%20Study%20of%20Surrogate%20Techniques%20for%20CNN%20Hyperparameter%20Optimization.pdf Mohd Aszemi, Nurshazlyn and M. Zakaria, Nordin and Paneer Selvam, Dhanapal Durai Dominic (2022) Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization. In: New Frontiers in Communication and Intelligent Systems. Computing & Intelligent Systems, SCRS, India, pp. 463-473. ISBN 978-81-95502-00-4 https://www.publications.scrs.in/chapter/978-81-95502-00-4/48 https://doi.org/10.52458/978-81-95502-00-4-48 https://doi.org/10.52458/978-81-95502-00-4-48 |
institution |
Universiti Teknologi Petronas |
building |
UTP Resource Centre |
collection |
Institutional Repository |
continent |
Asia |
country |
Malaysia |
content_provider |
Universiti Teknologi Petronas |
content_source |
UTP Electronic and Digitized Intellectual Asset |
url_provider |
http://utpedia.utp.edu.my/ |
language |
English |
topic |
T Technology (General) |
spellingShingle |
T Technology (General) Mohd Aszemi, Nurshazlyn M. Zakaria, Nordin Paneer Selvam, Dhanapal Durai Dominic Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization |
description |
Optimizing hyper parameters in Convolutional Neural networks is a tedious process for many researchers and practitioners. It requires a high degree of expertise or experience to optimise the hyper parameters, and manual optimisation is likely to be biased. To date, methods or approaches to automate hyper parameter optimization include grid search, random search, and Genetic Algorithms (GAs). However, evaluating large number of sample points in the hyperparameter configuration space, as is typically required by these methods, is computationally expensive process. Hence, the objective of this paper is to explore regression as a surrogate technique in CNN hyperparameter optimisation. Performance in terms of accuracy, error rate, training time and coefficient of determination (R2) are evaluated and recorded. Although there is no significant performance difference between the resulting optimized Deep Learning and state-of-the-art on CIFAR-10 datasets, using
regression as a surrogate technique for CNN hyperparameter optimization contributes to minimising the time taken for the optimization process, a benefit which has not been fully explored in the literature to the best of the author’s knowledge. |
format |
Book Section |
author |
Mohd Aszemi, Nurshazlyn M. Zakaria, Nordin Paneer Selvam, Dhanapal Durai Dominic |
author_facet |
Mohd Aszemi, Nurshazlyn M. Zakaria, Nordin Paneer Selvam, Dhanapal Durai Dominic |
author_sort |
Mohd Aszemi, Nurshazlyn |
title |
Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization |
title_short |
Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization |
title_full |
Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization |
title_fullStr |
Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization |
title_full_unstemmed |
Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization |
title_sort |
comparative study of surrogate techniques for cnn hyperparameter optimization |
publisher |
Computing & Intelligent Systems, SCRS |
publishDate |
2022 |
url |
http://utpedia.utp.edu.my/id/eprint/24082/1/Comparative%20Study%20of%20Surrogate%20Techniques%20for%20CNN%20Hyperparameter%20Optimization.pdf https://doi.org/10.52458/978-81-95502-00-4-48 http://utpedia.utp.edu.my/id/eprint/24082/ https://www.publications.scrs.in/chapter/978-81-95502-00-4/48 https://doi.org/10.52458/978-81-95502-00-4-48 |
_version_ |
1768010132250165248 |
score |
13.214268 |