Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization

Optimizing hyper parameters in Convolutional Neural networks is a tedious process for many researchers and practitioners. It requires a high degree of expertise or experience to optimise the hyper parameters, and manual optimisation is likely to be biased. To date, methods or approaches to automate...

Full description

Saved in:
Bibliographic Details
Main Authors: Mohd Aszemi, Nurshazlyn, M. Zakaria, Nordin, Paneer Selvam, Dhanapal Durai Dominic
Format: Book Section
Language:English
Published: Computing & Intelligent Systems, SCRS 2022
Subjects:
Online Access:http://utpedia.utp.edu.my/id/eprint/24082/1/Comparative%20Study%20of%20Surrogate%20Techniques%20for%20CNN%20Hyperparameter%20Optimization.pdf
https://doi.org/10.52458/978-81-95502-00-4-48
http://utpedia.utp.edu.my/id/eprint/24082/
https://www.publications.scrs.in/chapter/978-81-95502-00-4/48
https://doi.org/10.52458/978-81-95502-00-4-48
Tags: Add Tag
No Tags, Be the first to tag this record!
id oai:utpedia.utp.edu.my:24082
record_format eprints
spelling oai:utpedia.utp.edu.my:240822023-05-15T07:44:34Z http://utpedia.utp.edu.my/id/eprint/24082/ Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization Mohd Aszemi, Nurshazlyn M. Zakaria, Nordin Paneer Selvam, Dhanapal Durai Dominic T Technology (General) Optimizing hyper parameters in Convolutional Neural networks is a tedious process for many researchers and practitioners. It requires a high degree of expertise or experience to optimise the hyper parameters, and manual optimisation is likely to be biased. To date, methods or approaches to automate hyper parameter optimization include grid search, random search, and Genetic Algorithms (GAs). However, evaluating large number of sample points in the hyperparameter configuration space, as is typically required by these methods, is computationally expensive process. Hence, the objective of this paper is to explore regression as a surrogate technique in CNN hyperparameter optimisation. Performance in terms of accuracy, error rate, training time and coefficient of determination (R2) are evaluated and recorded. Although there is no significant performance difference between the resulting optimized Deep Learning and state-of-the-art on CIFAR-10 datasets, using regression as a surrogate technique for CNN hyperparameter optimization contributes to minimising the time taken for the optimization process, a benefit which has not been fully explored in the literature to the best of the author’s knowledge. Computing & Intelligent Systems, SCRS 2022-05-06 Book Section PeerReviewed text en http://utpedia.utp.edu.my/id/eprint/24082/1/Comparative%20Study%20of%20Surrogate%20Techniques%20for%20CNN%20Hyperparameter%20Optimization.pdf Mohd Aszemi, Nurshazlyn and M. Zakaria, Nordin and Paneer Selvam, Dhanapal Durai Dominic (2022) Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization. In: New Frontiers in Communication and Intelligent Systems. Computing & Intelligent Systems, SCRS, India, pp. 463-473. ISBN 978-81-95502-00-4 https://www.publications.scrs.in/chapter/978-81-95502-00-4/48 https://doi.org/10.52458/978-81-95502-00-4-48 https://doi.org/10.52458/978-81-95502-00-4-48
institution Universiti Teknologi Petronas
building UTP Resource Centre
collection Institutional Repository
continent Asia
country Malaysia
content_provider Universiti Teknologi Petronas
content_source UTP Electronic and Digitized Intellectual Asset
url_provider http://utpedia.utp.edu.my/
language English
topic T Technology (General)
spellingShingle T Technology (General)
Mohd Aszemi, Nurshazlyn
M. Zakaria, Nordin
Paneer Selvam, Dhanapal Durai Dominic
Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization
description Optimizing hyper parameters in Convolutional Neural networks is a tedious process for many researchers and practitioners. It requires a high degree of expertise or experience to optimise the hyper parameters, and manual optimisation is likely to be biased. To date, methods or approaches to automate hyper parameter optimization include grid search, random search, and Genetic Algorithms (GAs). However, evaluating large number of sample points in the hyperparameter configuration space, as is typically required by these methods, is computationally expensive process. Hence, the objective of this paper is to explore regression as a surrogate technique in CNN hyperparameter optimisation. Performance in terms of accuracy, error rate, training time and coefficient of determination (R2) are evaluated and recorded. Although there is no significant performance difference between the resulting optimized Deep Learning and state-of-the-art on CIFAR-10 datasets, using regression as a surrogate technique for CNN hyperparameter optimization contributes to minimising the time taken for the optimization process, a benefit which has not been fully explored in the literature to the best of the author’s knowledge.
format Book Section
author Mohd Aszemi, Nurshazlyn
M. Zakaria, Nordin
Paneer Selvam, Dhanapal Durai Dominic
author_facet Mohd Aszemi, Nurshazlyn
M. Zakaria, Nordin
Paneer Selvam, Dhanapal Durai Dominic
author_sort Mohd Aszemi, Nurshazlyn
title Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization
title_short Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization
title_full Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization
title_fullStr Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization
title_full_unstemmed Comparative Study of Surrogate Techniques for CNN Hyperparameter Optimization
title_sort comparative study of surrogate techniques for cnn hyperparameter optimization
publisher Computing & Intelligent Systems, SCRS
publishDate 2022
url http://utpedia.utp.edu.my/id/eprint/24082/1/Comparative%20Study%20of%20Surrogate%20Techniques%20for%20CNN%20Hyperparameter%20Optimization.pdf
https://doi.org/10.52458/978-81-95502-00-4-48
http://utpedia.utp.edu.my/id/eprint/24082/
https://www.publications.scrs.in/chapter/978-81-95502-00-4/48
https://doi.org/10.52458/978-81-95502-00-4-48
_version_ 1768010132250165248
score 13.214268