AI Chat Paper
Note: Please note that the following content is generated by AMiner AI. SciOpen does not take any responsibility related to this content.
{{lang === 'zh_CN' ? '文章概述' : 'Summary'}}
{{lang === 'en_US' ? '中' : 'Eng'}}
Chat more with AI
PDF (2.4 MB)
Collect
Submit Manuscript AI Chat Paper
Show Outline
Outline
Show full outline
Hide outline
Outline
Show full outline
Hide outline
Open Access | Online First

Image-Text Sentiment Analysis Using Modality Adaption and Multi-Broad Learning

School of Computer Science and Engineering, Guangdong Ocean University, Yangjiang 529500, China
School of Ocean Engineering and Technology, Sun Yat-sen University, Zhuhai 519082, China, and also with School of Computer Science and Engineering, Guangdong Ocean University, Yangjiang 529500, China
Institute of Language Learning and Artificial Intelligence, Guangdong University of Foreign Studies, Guangzhou 510006, China
School of English Education, Guangdong University of Foreign Studies, Guangzhou 510006, ChLina
Show Author Information

Abstract

Image-text sentiment analysis task has attracted increasing attention in recent years because of the surge of social media reviews in social networks. Although previous research works have made significant progress with feature fusion between image and text modalities, how to effectively obtain the intra-modality and inter-modality features is still an open research issue in image-text sentiment analysis. To address this problem, we propose a novel method called Modality Adaptation Multi-Broad Learning (MAMBL). Specifically, we take Vision Transformer (ViT) and Robustly optimized Bidirectional Encoder Representation from Transformers approach (RoBERTa) pre-training models to extract image and text features, respectively. Then, we adopt Multi-Layer Perceptron (MLP) unit to learn modality-invariant and modality-specific representations to provide a comprehensive view for understanding image-text data. Furthermore, we introduce two Dual Broad Learning (DBL) to fuse multi-modal features for sentiment classification. Extensive experiments have conducted on three benchmark image-text sentiment analysis datasets, namely MVSA-Single, MVSA-Multiple, and HFM. The experimental results demonstrate that our proposed method can achieve higher performance than the baseline models.

References

【1】
【1】
 
 
Tsinghua Science and Technology

{{item.num}}

Comments on this article

Go to comment

< Back to all reports

Review Status: {{reviewData.commendedNum}} Commended , {{reviewData.revisionRequiredNum}} Revision Required , {{reviewData.notCommendedNum}} Not Commended Under Peer Review

Review Comment

Close
Close
Cite this article:
Deng C, Yu J, Peng S, et al. Image-Text Sentiment Analysis Using Modality Adaption and Multi-Broad Learning. Tsinghua Science and Technology, 2026, https://doi.org/10.26599/TST.2025.9010152

613

Views

61

Downloads

0

Crossref

0

Web of Science

0

Scopus

0

CSCD

Received: 30 June 2025
Revised: 13 September 2025
Accepted: 25 September 2025
Published: 06 February 2026
© The author(s) 2026.

The articles published in this open access journal are distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/).