重庆大学学报2024,Vol.47Issue(4):139-156,18.DOI:10.11835/j.issn.1000.582X.2024.04.012
面向自动驾驶的多模态信息融合动态目标识别
Multimodal information fusion dynamic target recognition for autonomous driving
摘要
Abstract
A multi-modal information fusion based object recognition method for autonomous driving is proposed to address the vehicle and pedestrian detection challenge in autonomous driving environments.The method first improves ResNet50 network based on spatial attention mechanism and hybrid null convolution.The standard convolution is replaced by selective kernel convolution,which allows the network to dynamically adjust the size of the perceptual field according to the feature size.Then,the sawtooth hybrid null convolution is used to enable the network to capture multi-scale contextual information and improve the network feature extraction capability.The localization loss function in YOLOv3 is replaced with the GIoU loss function,which has better operability in practical applications.Finally,human-vehicle target classification and recognition algorithm based on two kinds of data fusion is proposed,which can improve the accuracy of the target detection.Experimental results show that compared with OFTNet,VoxelNet and FASTERRCNN,the mAP index can be improved by 0.05 during daytime and 0.09 in the evening,and the convergence effect is good.关键词
自动驾驶/ResNet50/YOLOv3/数据融合/注意力机制/损失函数Key words
autonomous driving/ResNet50/YOLOv3/data fusion/attention mechanism/loss function分类
通用工业技术引用本文复制引用
张明容,喻皓,吕辉,姜立标,李利平,卢磊..面向自动驾驶的多模态信息融合动态目标识别[J].重庆大学学报,2024,47(4):139-156,18.基金项目
国家自然科学基金资助项目(51975217).Supported by National Natural Science Foundation of China(51975217). (51975217)