Real-Time Multi-Task ADAS Implementation on Reconfigurable Heterogeneous MPSoC Architecture
Citation
TATAR, Güner, & Salih BAYAR. "Real-Time Multi-Task ADAS Implementation on Reconfigurable Heterogeneous MPSoC Architecture". IEEE Access, 4 (2023): 1-21.Abstract
The rapid adoption of Advanced Driver Assistance Systems (ADAS) in modern vehicles,
aiming to elevate driving safety and experience, necessitates the real-time processing of high-definition
video data. This requirement brings about considerable computational complexity and memory demands,
highlighting a critical research void for a design integrating high FPS throughput with optimal Mean
Average Precision (mAP) and Mean Intersection over Union (mIoU). Performance improvement at lower
costs, multi-tasking ability on a single hardware platform, and flawless incorporation into memoryconstrained
devices are also essential for boosting ADAS performance. Addressing these challenges,
this study proposes an ADAS multi-task learning hardware-software co-design approach underpinned
by the Kria KV260 Multi-Processor System-on-Chip Field Programmable Gate Array (MPSoC-FPGA)
platform. The approach facilitates efficient real-time execution of deep learning algorithms specific to ADAS
applications. Utilizing the BDD100K, KITTI, and CityScapes datasets, our ADAS multi-task learning
system endeavours to provide accurate and efficient multi-object detection, segmentation, and lane and
drivable area detection in road images. The system deploys a segmentation-based object detection strategy,
using a ResNet-18 backbone encoder and a Single Shot Detector architecture, coupled with quantizationaware
training to augment inference performance without compromising accuracy. The ADAS multi-task
learning offers customization options for various ADAS applications and can be further optimized for
increased precision and reduced memory usage. Experimental results showcase the system’s capability to
perform real-time multi-class object detection, segmentation, line detection, and drivable area detection on
road images at approximately 25.4 FPS using a 1920x1080p Full HD camera. Impressively, the quantized
model has demonstrated a 51% mAP for object detection, 56.62% mIoU for image segmentation, 43.86%
mIoU for line detection, and 81.56% IoU for drivable area identification, reinforcing its high efficacy
and precision. The findings underscore that the proposed ADAS multi-task learning system is a practical,
reliable, and effective solution for real-world applications.