Utilize este identificador para referenciar este registo: https://hdl.handle.net/1822/79802

TítuloCustomizable FPGA-based hardware accelerator for standard convolution processes empowered with quantization applied to LiDAR data
Autor(es)Silva, João Pedro Duarte
Pereira, Pedro Miguel Coelho
Machado, Rui
Névoa, Rafael
Melo-Pinto, Pedro
Fernandes, Duarte
Palavras-chaveconvolutional neural network (CNN)
hardware accelerator
field-programmable gate array (FPGA)
light detection and ranging (LiDAR)
quantization
object detection
Data11-Mar-2022
EditoraMultidisciplinary Digital Publishing Institute
RevistaSensors
CitaçãoSilva, J.; Pereira, P.; Machado, R.; Névoa, R.; Melo-Pinto, P.; Fernandes, D. Customizable FPGA-Based Hardware Accelerator for Standard Convolution Processes Empowered with Quantization Applied to LiDAR Data. Sensors 2022, 22, 2184. https://doi.org/10.3390/s22062184
Resumo(s)In recent years there has been an increase in the number of research and developments in deep learning solutions for object detection applied to driverless vehicles. This application benefited from the growing trend felt in innovative perception solutions, such as LiDAR sensors. Currently, this is the preferred device to accomplish those tasks in autonomous vehicles. There is a broad variety of research works on models based on point clouds, standing out for being efficient and robust in their intended tasks, but they are also characterized by requiring point cloud processing times greater than the minimum required, given the risky nature of the application. This research work aims to provide a design and implementation of a hardware IP optimized for computing convolutions, rectified linear unit (ReLU), padding, and max pooling. This engine was designed to enable the configuration of features such as varying the size of the feature map, filter size, stride, number of inputs, number of filters, and the number of hardware resources required for a specific convolution. Performance results show that by resorting to parallelism and quantization approach, the proposed solution could reduce the amount of logical FPGA resources by 40 to 50%, enhancing the processing time by 50% while maintaining the deep learning operation accuracy.
TipoArtigo
URIhttps://hdl.handle.net/1822/79802
DOI10.3390/s22062184
ISSN1424-8220
e-ISSN1424-8220
Versão da editorahttps://www.mdpi.com/1424-8220/22/6/2184
Arbitragem científicayes
AcessoAcesso aberto
Aparece nas coleções:CAlg - Artigos em revistas internacionais / Papers in international journals

Ficheiros deste registo:
Ficheiro Descrição TamanhoFormato 
sensors-22-02184.pdf7,1 MBAdobe PDFVer/Abrir

Partilhe no FacebookPartilhe no TwitterPartilhe no DeliciousPartilhe no LinkedInPartilhe no DiggAdicionar ao Google BookmarksPartilhe no MySpacePartilhe no Orkut
Exporte no formato BibTex mendeley Exporte no formato Endnote Adicione ao seu ORCID