整体与局部相互感知的图网络时序动作检测OACSTPCD
Temporal action detection via overall-local-aware graph network
针对目前动作检测与定位方法未综合利用整体与局部相互感知的时空关系信息、不利于提升动作检测与定位性能的问题,提出整体与局部相互感知的图网络时序动作检测方法.该方法综合利用各动作提案的特征相似性和时序重叠度构建整体关系图推理子网络,通过学习获得提案,该提案包含更丰富的整体时空特征表示;利用提案发生的时间偏序关系,构建局部关系图推理子网络,该子网络包含多个级别三体相似图和三体互补图的结构,通过学习获得不同时间尺度下提案的局部关系信息;最后构成整体与局部关系相互感知的丰富特征表达,用于动作检测与定位.采用平均精度均值作为评价指标在2 个公开数据集(Thumos14 和ActivityNet1.3)上进行了试验.结果表明,与PGCN、G-TAD、TAL-Net、CDC等先进方法相比,文中方法能有效提高动作检测的性能.
To solve the problem that the comprehensive utilization of overall-local-aware spatio-temporal relationship information was not considered in current action prediction and localization and was not conducive to improving the performance of action detection and localization,a temporal action detection method based on overall-local-aware graph network was proposed.To obtain richer overall spatio-temporal feature representation of proposals,the feature similarity and temporal overlap of each action proposal was comprehensively exploited to construct the overall relation graph reasoning sub-network of proposals.To obtain local relation information of proposals under different time scales,the partial order relationship over time for the proposals was exploited,and the local relation graph reasoning sub-network was constructed,which consisted of multiple levels of three-body similar graphs and three-body complementary graphs.The rich overall-local aware features for the proposals were represented,which were used to predict and localize actions.The experiments were conducted on two public datasets of Thumos14 and ActivityNet1.3 and measured by the mean average precision metric.The results show that compared with the advanced methods of PGCN,G-TAD,TAL-Net and CDC,the proposed method can effectively improve the performance of action detection.
黄金钾;詹永照;赵逸飞
江苏大学 计算机科学与通信工程学院, 江苏 镇江 212013
计算机与自动化
计算机视觉时序动作检测注意力机制整体与局部相互感知图网络时空特征表达
computer visiontemporal action detectionattention mechanismoverall-local-awaregraph networkspatio-temporal feature representation
《江苏大学学报(自然科学版)》 2024 (001)
67-76 / 10
国家自然科学基金资助项目(61672268)
评论