红外技术2025,Vol.47Issue(4):468-474,7.
基于特征交互与自适应分组融合的多模态目标检测
Multimodal Object Detection Based on Feature Interaction and Adaptive Grouping Fusion
摘要
Abstract
To improve the performance of object detection methods in complex scenes,a multimodal object detection model based on feature interaction and adaptive grouping fusion is proposed by combining deep learning algorithms with multimodal information fusion technology.The model uses infrared and visible object images as inputs,constructs a symmetrical dual-branch feature extraction structure based on the PP-LCNet network,and introduces a feature interaction module to ensure complementary information between different modal object features during the extraction process.Secondly,a binary grouping attention mechanism was designed.Global pooling combined with the sign function was used to group the output features of the interaction module into their respective object categories,and spatial attention mechanisms were used to enhance the object information in each group of features.Finally,based on the group-enhanced features,similar feature groups at different scales were extracted,and multi-scale fusion was carried out through adaptive weighting from deep to shallow.Object prediction was then achieved based on the fused features at each scale.The experimental results show that the proposed method significantly improves multimodal feature interaction,key feature enhancement,and multi-scale fusion.Moreover,in complex scenarios,the model exhibits higher robustness and can be better applied to different scenarios.关键词
多模态/目标检测/特征交互/二值化分组/自适应融合Key words
multimodal/object detection/feature interaction/binary grouping/adaptive fusion分类
信息技术与安全科学引用本文复制引用
叶志晖,武健,赵晓忠,王文娟,邵新光..基于特征交互与自适应分组融合的多模态目标检测[J].红外技术,2025,47(4):468-474,7.基金项目
国家自然科学基金(62002320). (62002320)