| 注册
首页|期刊导航|计算机应用研究|基于超图学习与成对跨模态融合的多模态对话情绪识别

基于超图学习与成对跨模态融合的多模态对话情绪识别

李尚往 缪裕青 刘同来 张万桢 周明

计算机应用研究2026,Vol.43Issue(2):361-368,8.
计算机应用研究2026,Vol.43Issue(2):361-368,8.DOI:10.19734/j.issn.1001-3695.2025.07.0233

基于超图学习与成对跨模态融合的多模态对话情绪识别

Multimodal emotion recognition in conversation based on hypergraph learning and pairwise cross-modal fusion

李尚往 1缪裕青 2刘同来 3张万桢 3周明4

作者信息

  • 1. 桂林电子科技大学计算机与信息安全学院,广西桂林 541004
  • 2. 桂林电子科技大学计算机与信息安全学院,广西桂林 541004||桂林电子科技大学广西图像图形与智能处理重点实验室,广西桂林 541004
  • 3. 仲恺农业工程学院人工智能学院,广州 510225
  • 4. 桂林海威科技股份有限公司,广西 桂林 541004
  • 折叠

摘要

Abstract

To address issues such as insufficient utilization of interaction information between modalities and multivariate dia-logue relations in current multimodal emotion recognition in conversation models,this paper proposed a multimodal emotion recognition in conversation model based on hypergraph learning and pairwise cross-modal fusion.In the hypergraph learning module of the model,it took discourse representations as nodes,and designed two different types of hyperedges containing multimodal and temporal information to form a hypergraph.It used hypergraph convolution to capture multivariate dialogue re-lations between speakers.Meanwhile,this paper proposed a dual-stream gated attention network to dynamically adjust node features and reduce information redundancy.In the pairwise cross-modal fusion module,it used each modality as a baseline feature.And based on the cross-modal attention mechanism,it was repeatedly reinforced with other modal features to excavate deep interaction information between pairwise modalities and enhance cross-modal feature representation.Experimental results show that on the IEMOCAP and CMU-MOSEI datasets,the accuracy and weighted average F1 score of the proposed model are better than those of multiple comparison models,fully verifying the effectiveness of the model.

关键词

对话情绪识别/超图/跨模态融合/双流门控注意力网络/Transformer

Key words

emotion recognition in conversation/hypergraph/cross-modal fusion/dual-stream gated attention network/Transformer

分类

信息技术与安全科学

引用本文复制引用

李尚往,缪裕青,刘同来,张万桢,周明..基于超图学习与成对跨模态融合的多模态对话情绪识别[J].计算机应用研究,2026,43(2):361-368,8.

基金项目

国家自然科学基金资助项目(62366010,62366011) (62366010,62366011)

广东省自然科学基金资助项目(2023A1515011230) (2023A1515011230)

广东省哲学社会科学规划专项项目(GD25CW04) (GD25CW04)

桂林电子科技大学研究生教育创新计划资助项目(2025YCXS076) (2025YCXS076)

计算机应用研究

1001-3695

访问量0
|
下载量0
段落导航相关论文