融合改进图卷积的跨模态检索OA北大核心CSTPCD
Cross-Modal Retrieval with Improved Graph Convolution
针对现有跨模态检索在公共子空间度量时难以充分挖掘模态内局部一致性的问题,提出了一种融合改进图卷积的跨模态检索方法.为了提升各模态内的局部一致性,以单个完整样本为节点构建模态图,充分挖掘特征间的交互信息;为了解决图卷积网络只能做浅层学习的问题,采用在每一层图卷积添加初始残差链接和权重恒等映射的方法来缓解此现象;为了通过高低阶邻居信息共同更新中心节点特征,提出减少邻居节点、增加图卷积网络层数的改进;为了学习高度局部一致且语义一致的公共表征,共享公共表征学习层权重,并联合优化公共子空间中模态内的语义约束和模态间的模态不变约束.实验结果表明,在Wikipedia和Pascal sentence这两个跨模态数据集上,不同检索任务的平均mAP值比11种现有方法分别提升了2.2%~42.1%和3.0%~54.0%.
Aiming at the problem that existing image text cross-modal retrieval is difficult to fully exploit the local consis-tency in the mode in the common subspace,a cross-modal retrieval method based on improved graph convolution is pro-posed.In order to improve the local consistency within each mode,the modal diagram is constructed with a single sample as a node,fully mining the interactive information between features.In order to solve the problem that graph convolution network can only do shallow learning,the method of adding initial residual link and weight identity map in each layer of graph convolution is adopted to alleviate this phenomenon.In order to jointly update the central node features through higher-order and lower-order neighbor information,an improvement is proposed to reduce neighbor nodes and increase the number of layers in graph convolution network.In order to learn highly locally consistent and semantically consistent public representation,it shares the weights of common representation learning layer,and jointly optimizes the semantic constraints within the modes and the modal invariant constraints between modes in the common subspace.The experimen-tal results show that on the two cross-modal data sets of Wikipedia and Pascal sentence,the average mAP values of differ-ent retrieval tasks are 2.2%~42.1%and 3.0%~54.0%higher than the 11 existing methods.
张宏图;化春键;蒋毅;俞建峰;陈莹
江南大学 机械工程学院,江苏 无锡 214122||江苏省食品先进制造装备技术重点实验室,江苏 无锡 214122江南大学 物联网工程学院,江苏 无锡 214122
计算机与自动化
图卷积网络跨模态检索初始残差连接恒等映射邻接矩阵
graph convolution networkcross-modal retrievalinitial residual connectionidentity mappingadjacency matrix
《计算机工程与应用》 2024 (011)
95-104 / 10
国家自然科学基金(62173160).
评论