학술논문

Overtaking Mechanisms Based on Augmented Intelligence for Autonomous Driving: Data Sets, Methods, and Challenges
Document Type
Periodical
Source
IEEE Internet of Things Journal IEEE Internet Things J. Internet of Things Journal, IEEE. 11(10):17911-17933 May, 2024
Subject
Computing and Processing
Communication, Networking and Broadcast Technologies
Automobiles
Roads
Autonomous automobiles
Pedestrians
Accidents
Decision making
Task analysis
Augmented intelligence
autonomous driving
computer vision
overtaking mechanism
Language
ISSN
2327-4662
2372-2541
Abstract
The field of autonomous driving research has made significant strides toward achieving full automation, endowing vehicles with self-awareness and independent decision making. However, integrating automation into vehicular operations presents formidable challenges, especially as these vehicles must seamlessly navigate public roads alongside other cars and pedestrians. An intriguing yet relatively underexplored domain within autonomous driving is overtaking. Overtaking involves a dynamic interplay of complex tasks, including precise steering and speed control, rendering it one of the most intricate operations for implementing augmented intelligence driving technologies. Surprisingly, the overtaking of autonomous vehicles (AVs) remains largely uncharted territory in the context of augmented intelligence for autonomous systems. This void in knowledge beckons researchers to embark on explorations and investigations in this nascent field. Our review paper systematically synthesises overtaking methodologies hinging on computer vision techniques tailored for augmented intelligence autonomous driving scenarios in response to this pressing need. Our analysis encompasses an array of domains central to overtaking in augmented intelligence AVs, encompassing Object Detection, Lane/Line Detection, Depth Estimation, Obstacle Detection, Segmentation, and Pedestrian Detection. We meticulously analyze each domain using well-established multimodal data sets. We assess different models’ performance across various parameters by employing graphical structures, enabling visual comparative analyses. In object detection, YOLOv4 achieves a top performance with 0.90 mAP on the BDD100K data set. For lane detection, CLRNET excels with the highest F1 score of around 0.96 on the LLAMAS data set. ViT-Adapter-L leads in segmentation tasks, boasting an impressive mIoU score of 83 on Cityscapes. The Hierarchical Model achieves a superior mean Average Precision of 0.90 in road sign detection on the Tsinghua-Tencent data set. Steering angle computation sees InterFuser as the standout, achieving the highest driving score of approximately 74. This article’s primary contributions include a comprehensive assessment of diverse models for each Multimodal data set, aiding future research in this evolving domain.