| 注册
首页|期刊导航|数据采集与处理|基于GCN和目标视觉特征增强的多模态方面级情感分析

基于GCN和目标视觉特征增强的多模态方面级情感分析

赵雪峰 柏长泽 狄恒西 仲兆满 仲晓敏

数据采集与处理2025,Vol.40Issue(5):1177-1192,16.
数据采集与处理2025,Vol.40Issue(5):1177-1192,16.DOI:10.16337/j.1004-9037.2025.05.006

基于GCN和目标视觉特征增强的多模态方面级情感分析

Multimodal Aspect-Level Sentiment Analysis Based on GCN and Target Visual Feature Enhancement

赵雪峰 1柏长泽 1狄恒西 1仲兆满 1仲晓敏1

作者信息

  • 1. 江苏海洋大学计算机工程学院,连云港 222005
  • 折叠

摘要

Abstract

Multimodal aspect-level sentiment analysis aims to integrate graphic modal data to accurately predict the emotional polarity of aspect words.However,the existing methods still have significant limitations in accurately locating text-related image region features and effectively processing the information interaction between modalities.At the same time,the understanding of context information within modalities is biased,which leads to additional noise.In order to solve the above problems,a multi-modal aspect-level sentiment analysis model based on graph convolutional network and target visual feature enhancement(GCN-TVFE)is proposed.First of all,this paper uses the contrastive language-image pre-training(CLIP)model to process text,aspect words,and image data.By calculating the similarity between text and image and the similarity between aspect words and image,and then combining these two similarities,the quantitative evaluation of the matching degree between text and image and the matching degree of aspect words and image is realized.Then,the Faster R-CNN model is used to quickly and accurately identify and locate the target region in the image,which further enhances the ability of the model to extract image features related to text.Secondly,through the GCN network,the text graph structure is constructed by using the dependency syntactic relationship between texts,and the image graph structure is generated by the K-nearest neighbor(KNN)algorithm,to dig the feature information in the mode deeply.Finally,the multi-layer and multi-modal interactive attention mechanism is used to effectively capture the correlation information between aspect words and text,and between target visual features and image-generated text description features,which significantly reduces noise interference and enhances feature interaction between modes.Experimental results show that the model proposed in this paper has superior comprehensive performance on the public datasets Twitter-2015 and Twitter-2017,which verifies the effectiveness of the model in the field of multimodal sentiment analysis.

关键词

多模态方面级情感分析/目标视觉特征/依存句法关系/KNN算法/多模态交互注意力机制

Key words

multimodal aspect-level sentiment analysis/target visual features/dependency syntactic relation/KNN algorithm/multimodal interactive attention mechanism

分类

信息技术与安全科学

引用本文复制引用

赵雪峰,柏长泽,狄恒西,仲兆满,仲晓敏..基于GCN和目标视觉特征增强的多模态方面级情感分析[J].数据采集与处理,2025,40(5):1177-1192,16.

基金项目

国家自然科学基金(72174079) (72174079)

江苏省"青蓝工程"优秀教学团队项目(2022-29). (2022-29)

数据采集与处理

OA北大核心

1004-9037

访问量0
|
下载量0
段落导航相关论文