A significant amount of Deep Learning research deals with the reduction of network complexity. In most scenarios the preservation of very high performance has priority over size reduction. However, when dealing with embedded systems, the limited amount of resources forces a switch in perspective. In fact, being able to dramatically reduce complexity could be a stronger requisite for overall feasibility than excellent performance. In this paper we propose a simple to implement yet effective method to largely reduce the size of Convolutional Neural Networks with minimal impact on their performance. The key idea is to assess the relevance of each kernel with respect to a representative dataset by computing the output of its activation function and to trim them accordingly. The resulting network becomes small enough to be adopted on embedded hardware, such as smart cameras or lightweight edge processing units. In order to assess the capability of our method with respect to real-world scenarios, we adopted it to shrink two different pre-trained networks to be hosted on general purpose low-end FPGA hardware to be found in embedded cameras. Our experiments demonstrated both the overall feasibility of the method and its superior performance when compared with similar size-reducing techniques introduced in recent literature.
A Relevance-Based CNN Trimming Method for Low-Resources Embedded Vision
Ressi D.
;Pistellato M.;Albarelli A.;Bergamasco F.
2022-01-01
Abstract
A significant amount of Deep Learning research deals with the reduction of network complexity. In most scenarios the preservation of very high performance has priority over size reduction. However, when dealing with embedded systems, the limited amount of resources forces a switch in perspective. In fact, being able to dramatically reduce complexity could be a stronger requisite for overall feasibility than excellent performance. In this paper we propose a simple to implement yet effective method to largely reduce the size of Convolutional Neural Networks with minimal impact on their performance. The key idea is to assess the relevance of each kernel with respect to a representative dataset by computing the output of its activation function and to trim them accordingly. The resulting network becomes small enough to be adopted on embedded hardware, such as smart cameras or lightweight edge processing units. In order to assess the capability of our method with respect to real-world scenarios, we adopted it to shrink two different pre-trained networks to be hosted on general purpose low-end FPGA hardware to be found in embedded cameras. Our experiments demonstrated both the overall feasibility of the method and its superior performance when compared with similar size-reducing techniques introduced in recent literature.File | Dimensione | Formato | |
---|---|---|---|
A_Relevance_Based_CNN_Trimming_Method_for_Low_Resources_Embedded_Vision.pdf
accesso aperto
Tipologia:
Documento in Pre-print
Licenza:
Accesso libero (no vincoli)
Dimensione
494.7 kB
Formato
Adobe PDF
|
494.7 kB | Adobe PDF | Visualizza/Apri |
I documenti in ARCA sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.