High performance accelerators for deep neural networks: A review
The availability of huge structured and unstructured data, advanced highly dense memory and high performance computing machines have provided a strong push for the development in artificial intelligence (AI) and machine learning (ML) domains. AI and machine learning has rekindled the hope of efficie...
Saved in:
Main Authors: | , , , , , , |
---|---|
Format: | Article |
Published: |
John Wiley and Sons Inc
2022
|
Subjects: | |
Online Access: | http://eprints.utm.my/103982/ http://dx.doi.org/10.1111/exsy.12831 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
id |
my.utm.103982 |
---|---|
record_format |
eprints |
spelling |
my.utm.1039822024-01-09T00:40:02Z http://eprints.utm.my/103982/ High performance accelerators for deep neural networks: A review Akhoon, Mohd. Saqib Suandi, Shahrel A. Alshahrani, Abdullah H. Y. Saad, Abdul-Malik Albogamy, Fahad R. Abdullah, Mohd. Zaid Loan, Sajad A. TK Electrical engineering. Electronics Nuclear engineering The availability of huge structured and unstructured data, advanced highly dense memory and high performance computing machines have provided a strong push for the development in artificial intelligence (AI) and machine learning (ML) domains. AI and machine learning has rekindled the hope of efficiently solving complex problems which was not possible in the recent past. The generation and availability of big-data is a strong driving force for the development of AI/ML applications, however, several challenges need to be addressed, like processing speed, memory requirement, high bandwidth, low latency memory access, and highly conductive and flexible connections between processing units and memory blocks. The conventional computing platforms are unable to address these issues with machine learning and AI. Deep neural networks (DNNs) are widely employed for machine learning and AI applications, like speech recognition, computer vison, robotics, and so forth, efficiently and accurately. However, accuracy is achieved at the cost of high computational complexity, sacrificing energy efficiency and throughput like performance measuring parameters along with high latency. To address the problems of latency, energy efficiency, complexity, power consumption, and so forth, a lot of state of the art DNN accelerators have been designed and implemented in the form of application specific integrated circuits (ASICs) and field programmable gate arrays (FPGAs). This work provides the state of the art of all these DNN accelerators which have been developed recently. Various DNN architectures, their computing units, emerging technologies used in improving the performance of DNN accelerators will be discussed. Finally, we will try to explore the scope for further improvement in these accelerator designs, various opportunities and challenges for the future research. John Wiley and Sons Inc 2022 Article PeerReviewed Akhoon, Mohd. Saqib and Suandi, Shahrel A. and Alshahrani, Abdullah and H. Y. Saad, Abdul-Malik and Albogamy, Fahad R. and Abdullah, Mohd. Zaid and Loan, Sajad A. (2022) High performance accelerators for deep neural networks: A review. Expert Systems, 39 (1). pp. 1-23. ISSN 0266-4720 http://dx.doi.org/10.1111/exsy.12831 DOI : 10.1111/exsy.12831 |
institution |
Universiti Teknologi Malaysia |
building |
UTM Library |
collection |
Institutional Repository |
continent |
Asia |
country |
Malaysia |
content_provider |
Universiti Teknologi Malaysia |
content_source |
UTM Institutional Repository |
url_provider |
http://eprints.utm.my/ |
topic |
TK Electrical engineering. Electronics Nuclear engineering |
spellingShingle |
TK Electrical engineering. Electronics Nuclear engineering Akhoon, Mohd. Saqib Suandi, Shahrel A. Alshahrani, Abdullah H. Y. Saad, Abdul-Malik Albogamy, Fahad R. Abdullah, Mohd. Zaid Loan, Sajad A. High performance accelerators for deep neural networks: A review |
description |
The availability of huge structured and unstructured data, advanced highly dense memory and high performance computing machines have provided a strong push for the development in artificial intelligence (AI) and machine learning (ML) domains. AI and machine learning has rekindled the hope of efficiently solving complex problems which was not possible in the recent past. The generation and availability of big-data is a strong driving force for the development of AI/ML applications, however, several challenges need to be addressed, like processing speed, memory requirement, high bandwidth, low latency memory access, and highly conductive and flexible connections between processing units and memory blocks. The conventional computing platforms are unable to address these issues with machine learning and AI. Deep neural networks (DNNs) are widely employed for machine learning and AI applications, like speech recognition, computer vison, robotics, and so forth, efficiently and accurately. However, accuracy is achieved at the cost of high computational complexity, sacrificing energy efficiency and throughput like performance measuring parameters along with high latency. To address the problems of latency, energy efficiency, complexity, power consumption, and so forth, a lot of state of the art DNN accelerators have been designed and implemented in the form of application specific integrated circuits (ASICs) and field programmable gate arrays (FPGAs). This work provides the state of the art of all these DNN accelerators which have been developed recently. Various DNN architectures, their computing units, emerging technologies used in improving the performance of DNN accelerators will be discussed. Finally, we will try to explore the scope for further improvement in these accelerator designs, various opportunities and challenges for the future research. |
format |
Article |
author |
Akhoon, Mohd. Saqib Suandi, Shahrel A. Alshahrani, Abdullah H. Y. Saad, Abdul-Malik Albogamy, Fahad R. Abdullah, Mohd. Zaid Loan, Sajad A. |
author_facet |
Akhoon, Mohd. Saqib Suandi, Shahrel A. Alshahrani, Abdullah H. Y. Saad, Abdul-Malik Albogamy, Fahad R. Abdullah, Mohd. Zaid Loan, Sajad A. |
author_sort |
Akhoon, Mohd. Saqib |
title |
High performance accelerators for deep neural networks: A review |
title_short |
High performance accelerators for deep neural networks: A review |
title_full |
High performance accelerators for deep neural networks: A review |
title_fullStr |
High performance accelerators for deep neural networks: A review |
title_full_unstemmed |
High performance accelerators for deep neural networks: A review |
title_sort |
high performance accelerators for deep neural networks: a review |
publisher |
John Wiley and Sons Inc |
publishDate |
2022 |
url |
http://eprints.utm.my/103982/ http://dx.doi.org/10.1111/exsy.12831 |
_version_ |
1787582541097598976 |
score |
13.209306 |