Journal article 496 views 152 downloads
Deep Neural Network Hardware Implementation Based on Stacked Sparse Autoencoder
IEEE Access, Pages: 1 - 1
Swansea University Author: Matheus Torquato
-
PDF | Version of Record
Download (9.01MB)
DOI (Published version): 10.1109/ACCESS.2019.2907261
Abstract
Deep learning techniques have been gaining prominence in the research world in the past years, however, the deep learning algorithms have high computational cost, making them hard to be used to several commercial applications. On the other hand, new alternatives have been studied and some methodolog...
Published in: | IEEE Access |
---|---|
ISSN: | 2169-3536 |
Published: |
2019
|
Online Access: |
Check full text
|
URI: | https://cronfa.swan.ac.uk/Record/cronfa49874 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
first_indexed |
2019-04-04T16:40:50Z |
---|---|
last_indexed |
2019-07-17T15:33:43Z |
id |
cronfa49874 |
recordtype |
SURis |
fullrecord |
<?xml version="1.0"?><rfc1807><datestamp>2019-07-17T09:51:27.6797438</datestamp><bib-version>v2</bib-version><id>49874</id><entry>2019-04-03</entry><title>Deep Neural Network Hardware Implementation Based on Stacked Sparse Autoencoder</title><swanseaauthors><author><sid>7a053c668886b4642286baed36fdba90</sid><ORCID>0000-0001-6356-3538</ORCID><firstname>Matheus</firstname><surname>Torquato</surname><name>Matheus Torquato</name><active>true</active><ethesisStudent>false</ethesisStudent></author></swanseaauthors><date>2019-04-03</date><deptcode>SCS</deptcode><abstract>Deep learning techniques have been gaining prominence in the research world in the past years, however, the deep learning algorithms have high computational cost, making them hard to be used to several commercial applications. On the other hand, new alternatives have been studied and some methodologies focusing on accelerating complex algorithms including those based on reconfigurable hardware has been showing significant results. Therefore, the objective of this work is to propose a neural network hardware implementation to be used in deep learning applications. The implementation was developed on a Field Programmable Gate Array (FPGA) and supports Deep Neural Network (DNN) trained with the Stacked Sparse Autoencoder (SSAE) technique. In order to allow DNNs with several inputs and layers on the FPGA, the systolic array technique was used in the entire architecture. Details regarding the designed implementation were evidenced, as well as the hardware area occupation in and the processing time for two different implementations. The results showed that both implementations achieved high throughput enabling Deep Learning techniques to be applied for problems with large data amounts.</abstract><type>Journal Article</type><journal>IEEE Access</journal><paginationStart>1</paginationStart><paginationEnd>1</paginationEnd><publisher/><issnElectronic>2169-3536</issnElectronic><keywords/><publishedDay>31</publishedDay><publishedMonth>12</publishedMonth><publishedYear>2019</publishedYear><publishedDate>2019-12-31</publishedDate><doi>10.1109/ACCESS.2019.2907261</doi><url/><notes/><college>COLLEGE NANME</college><department>Computer Science</department><CollegeCode>COLLEGE CODE</CollegeCode><DepartmentCode>SCS</DepartmentCode><institution>Swansea University</institution><apcterm/><lastEdited>2019-07-17T09:51:27.6797438</lastEdited><Created>2019-04-03T11:14:25.0904999</Created><path><level id="1">Faculty of Science and Engineering</level><level id="2">School of Engineering and Applied Sciences - Uncategorised</level></path><authors><author><firstname>Maria G. F.</firstname><surname>Coutinho</surname><order>1</order></author><author><firstname>Matheus</firstname><surname>Torquato</surname><orcid>0000-0001-6356-3538</orcid><order>2</order></author><author><firstname>Marcelo A. C.</firstname><surname>Fernandes</surname><order>3</order></author></authors><documents><document><filename>0049874-03042019111711.pdf</filename><originalFilename>coutinho2019v2.pdf</originalFilename><uploaded>2019-04-03T11:17:11.6730000</uploaded><type>Output</type><contentLength>9766741</contentLength><contentType>application/pdf</contentType><version>Version of Record</version><cronfaStatus>true</cronfaStatus><embargoDate>2019-04-03T00:00:00.0000000</embargoDate><copyrightCorrect>true</copyrightCorrect><language>eng</language></document></documents><OutputDurs/></rfc1807> |
spelling |
2019-07-17T09:51:27.6797438 v2 49874 2019-04-03 Deep Neural Network Hardware Implementation Based on Stacked Sparse Autoencoder 7a053c668886b4642286baed36fdba90 0000-0001-6356-3538 Matheus Torquato Matheus Torquato true false 2019-04-03 SCS Deep learning techniques have been gaining prominence in the research world in the past years, however, the deep learning algorithms have high computational cost, making them hard to be used to several commercial applications. On the other hand, new alternatives have been studied and some methodologies focusing on accelerating complex algorithms including those based on reconfigurable hardware has been showing significant results. Therefore, the objective of this work is to propose a neural network hardware implementation to be used in deep learning applications. The implementation was developed on a Field Programmable Gate Array (FPGA) and supports Deep Neural Network (DNN) trained with the Stacked Sparse Autoencoder (SSAE) technique. In order to allow DNNs with several inputs and layers on the FPGA, the systolic array technique was used in the entire architecture. Details regarding the designed implementation were evidenced, as well as the hardware area occupation in and the processing time for two different implementations. The results showed that both implementations achieved high throughput enabling Deep Learning techniques to be applied for problems with large data amounts. Journal Article IEEE Access 1 1 2169-3536 31 12 2019 2019-12-31 10.1109/ACCESS.2019.2907261 COLLEGE NANME Computer Science COLLEGE CODE SCS Swansea University 2019-07-17T09:51:27.6797438 2019-04-03T11:14:25.0904999 Faculty of Science and Engineering School of Engineering and Applied Sciences - Uncategorised Maria G. F. Coutinho 1 Matheus Torquato 0000-0001-6356-3538 2 Marcelo A. C. Fernandes 3 0049874-03042019111711.pdf coutinho2019v2.pdf 2019-04-03T11:17:11.6730000 Output 9766741 application/pdf Version of Record true 2019-04-03T00:00:00.0000000 true eng |
title |
Deep Neural Network Hardware Implementation Based on Stacked Sparse Autoencoder |
spellingShingle |
Deep Neural Network Hardware Implementation Based on Stacked Sparse Autoencoder Matheus Torquato |
title_short |
Deep Neural Network Hardware Implementation Based on Stacked Sparse Autoencoder |
title_full |
Deep Neural Network Hardware Implementation Based on Stacked Sparse Autoencoder |
title_fullStr |
Deep Neural Network Hardware Implementation Based on Stacked Sparse Autoencoder |
title_full_unstemmed |
Deep Neural Network Hardware Implementation Based on Stacked Sparse Autoencoder |
title_sort |
Deep Neural Network Hardware Implementation Based on Stacked Sparse Autoencoder |
author_id_str_mv |
7a053c668886b4642286baed36fdba90 |
author_id_fullname_str_mv |
7a053c668886b4642286baed36fdba90_***_Matheus Torquato |
author |
Matheus Torquato |
author2 |
Maria G. F. Coutinho Matheus Torquato Marcelo A. C. Fernandes |
format |
Journal article |
container_title |
IEEE Access |
container_start_page |
1 |
publishDate |
2019 |
institution |
Swansea University |
issn |
2169-3536 |
doi_str_mv |
10.1109/ACCESS.2019.2907261 |
college_str |
Faculty of Science and Engineering |
hierarchytype |
|
hierarchy_top_id |
facultyofscienceandengineering |
hierarchy_top_title |
Faculty of Science and Engineering |
hierarchy_parent_id |
facultyofscienceandengineering |
hierarchy_parent_title |
Faculty of Science and Engineering |
department_str |
School of Engineering and Applied Sciences - Uncategorised{{{_:::_}}}Faculty of Science and Engineering{{{_:::_}}}School of Engineering and Applied Sciences - Uncategorised |
document_store_str |
1 |
active_str |
0 |
description |
Deep learning techniques have been gaining prominence in the research world in the past years, however, the deep learning algorithms have high computational cost, making them hard to be used to several commercial applications. On the other hand, new alternatives have been studied and some methodologies focusing on accelerating complex algorithms including those based on reconfigurable hardware has been showing significant results. Therefore, the objective of this work is to propose a neural network hardware implementation to be used in deep learning applications. The implementation was developed on a Field Programmable Gate Array (FPGA) and supports Deep Neural Network (DNN) trained with the Stacked Sparse Autoencoder (SSAE) technique. In order to allow DNNs with several inputs and layers on the FPGA, the systolic array technique was used in the entire architecture. Details regarding the designed implementation were evidenced, as well as the hardware area occupation in and the processing time for two different implementations. The results showed that both implementations achieved high throughput enabling Deep Learning techniques to be applied for problems with large data amounts. |
published_date |
2019-12-31T04:01:07Z |
_version_ |
1763753137758273536 |
score |
11.037166 |