High performance accelerators for deep neural networks: A review

The availability of huge structured and unstructured data, advanced highly dense memory and high performance computing machines have provided a strong push for the development in artificial intelligence (AI) and machine learning (ML) domains. AI and machine learning has rekindled the hope of efficie...

Full description

Saved in:
Bibliographic Details
Main Authors: Akhoon, Mohd. Saqib, Suandi, Shahrel A., Alshahrani, Abdullah, H. Y. Saad, Abdul-Malik, Albogamy, Fahad R., Abdullah, Mohd. Zaid, Loan, Sajad A.
Format: Article
Published: John Wiley and Sons Inc 2022
Subjects:
Online Access:http://eprints.utm.my/103982/
http://dx.doi.org/10.1111/exsy.12831
Tags: Add Tag
No Tags, Be the first to tag this record!
id my.utm.103982
record_format eprints
spelling my.utm.1039822024-01-09T00:40:02Z http://eprints.utm.my/103982/ High performance accelerators for deep neural networks: A review Akhoon, Mohd. Saqib Suandi, Shahrel A. Alshahrani, Abdullah H. Y. Saad, Abdul-Malik Albogamy, Fahad R. Abdullah, Mohd. Zaid Loan, Sajad A. TK Electrical engineering. Electronics Nuclear engineering The availability of huge structured and unstructured data, advanced highly dense memory and high performance computing machines have provided a strong push for the development in artificial intelligence (AI) and machine learning (ML) domains. AI and machine learning has rekindled the hope of efficiently solving complex problems which was not possible in the recent past. The generation and availability of big-data is a strong driving force for the development of AI/ML applications, however, several challenges need to be addressed, like processing speed, memory requirement, high bandwidth, low latency memory access, and highly conductive and flexible connections between processing units and memory blocks. The conventional computing platforms are unable to address these issues with machine learning and AI. Deep neural networks (DNNs) are widely employed for machine learning and AI applications, like speech recognition, computer vison, robotics, and so forth, efficiently and accurately. However, accuracy is achieved at the cost of high computational complexity, sacrificing energy efficiency and throughput like performance measuring parameters along with high latency. To address the problems of latency, energy efficiency, complexity, power consumption, and so forth, a lot of state of the art DNN accelerators have been designed and implemented in the form of application specific integrated circuits (ASICs) and field programmable gate arrays (FPGAs). This work provides the state of the art of all these DNN accelerators which have been developed recently. Various DNN architectures, their computing units, emerging technologies used in improving the performance of DNN accelerators will be discussed. Finally, we will try to explore the scope for further improvement in these accelerator designs, various opportunities and challenges for the future research. John Wiley and Sons Inc 2022 Article PeerReviewed Akhoon, Mohd. Saqib and Suandi, Shahrel A. and Alshahrani, Abdullah and H. Y. Saad, Abdul-Malik and Albogamy, Fahad R. and Abdullah, Mohd. Zaid and Loan, Sajad A. (2022) High performance accelerators for deep neural networks: A review. Expert Systems, 39 (1). pp. 1-23. ISSN 0266-4720 http://dx.doi.org/10.1111/exsy.12831 DOI : 10.1111/exsy.12831
institution Universiti Teknologi Malaysia
building UTM Library
collection Institutional Repository
continent Asia
country Malaysia
content_provider Universiti Teknologi Malaysia
content_source UTM Institutional Repository
url_provider http://eprints.utm.my/
topic TK Electrical engineering. Electronics Nuclear engineering
spellingShingle TK Electrical engineering. Electronics Nuclear engineering
Akhoon, Mohd. Saqib
Suandi, Shahrel A.
Alshahrani, Abdullah
H. Y. Saad, Abdul-Malik
Albogamy, Fahad R.
Abdullah, Mohd. Zaid
Loan, Sajad A.
High performance accelerators for deep neural networks: A review
description The availability of huge structured and unstructured data, advanced highly dense memory and high performance computing machines have provided a strong push for the development in artificial intelligence (AI) and machine learning (ML) domains. AI and machine learning has rekindled the hope of efficiently solving complex problems which was not possible in the recent past. The generation and availability of big-data is a strong driving force for the development of AI/ML applications, however, several challenges need to be addressed, like processing speed, memory requirement, high bandwidth, low latency memory access, and highly conductive and flexible connections between processing units and memory blocks. The conventional computing platforms are unable to address these issues with machine learning and AI. Deep neural networks (DNNs) are widely employed for machine learning and AI applications, like speech recognition, computer vison, robotics, and so forth, efficiently and accurately. However, accuracy is achieved at the cost of high computational complexity, sacrificing energy efficiency and throughput like performance measuring parameters along with high latency. To address the problems of latency, energy efficiency, complexity, power consumption, and so forth, a lot of state of the art DNN accelerators have been designed and implemented in the form of application specific integrated circuits (ASICs) and field programmable gate arrays (FPGAs). This work provides the state of the art of all these DNN accelerators which have been developed recently. Various DNN architectures, their computing units, emerging technologies used in improving the performance of DNN accelerators will be discussed. Finally, we will try to explore the scope for further improvement in these accelerator designs, various opportunities and challenges for the future research.
format Article
author Akhoon, Mohd. Saqib
Suandi, Shahrel A.
Alshahrani, Abdullah
H. Y. Saad, Abdul-Malik
Albogamy, Fahad R.
Abdullah, Mohd. Zaid
Loan, Sajad A.
author_facet Akhoon, Mohd. Saqib
Suandi, Shahrel A.
Alshahrani, Abdullah
H. Y. Saad, Abdul-Malik
Albogamy, Fahad R.
Abdullah, Mohd. Zaid
Loan, Sajad A.
author_sort Akhoon, Mohd. Saqib
title High performance accelerators for deep neural networks: A review
title_short High performance accelerators for deep neural networks: A review
title_full High performance accelerators for deep neural networks: A review
title_fullStr High performance accelerators for deep neural networks: A review
title_full_unstemmed High performance accelerators for deep neural networks: A review
title_sort high performance accelerators for deep neural networks: a review
publisher John Wiley and Sons Inc
publishDate 2022
url http://eprints.utm.my/103982/
http://dx.doi.org/10.1111/exsy.12831
_version_ 1787582541097598976
score 13.209306