학술논문
A Mixed-Precision Transformer Accelerator With Vector Tiling Systolic Array for License Plate Recognition in Unconstrained Scenarios
Document Type
Periodical
Source
IEEE Transactions on Intelligent Transportation Systems IEEE Trans. Intell. Transport. Syst. Intelligent Transportation Systems, IEEE Transactions on. 25(12):20280-20294 Dec, 2024
Subject
Language
ISSN
1524-9050
1558-0016
1558-0016
Abstract
Power efficiency for license plate recognition (LPR) under unconstrained scenarios is a crucial factor in many edge-based real-world applications, e.g., autonomous vehicles whose power budget is limited. While a bulk of prior works have explored LPR approaches for unconstrained situations on CPU and GPU servers, these methods result in huge power dissipation, and are ineffective in challenging scenes. In this work, we present a mixed-precision (MP) Transformer hardware architecture to meet the requirements of power efficiency and satisfactory LPR accuracy in unconstrained scenarios, dedicated to implementing power-efficient edge accelerators for difficult LPR tasks. Firstly, MP-LPR Transformer model is proposed, where novel mixed-precision quantization and non-linear approximation techniques are tailored for low bit-width inference. Secondly, vector tiling systolic array Transformer architecture (VTSATA) is proposed with unique vector tiling systolic array (VTSA) and vector ALU (VALU) designs, where VTSA is used to accelerate matrix multiplications, and VALU is used to process non-linear operations. Finally, a FPGA accelerator prototype based on our approach is developed. Experimental results demonstrate that our hardware platform can reduce power consumption more than $20\times $ compared to GPU platforms, and for challenge subset on CCPD dataset, our method can further improve nearly 1% accuracy with respect to the state-of-the-art performance.