skip to main content
10.1145/3651671.3651755acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicmlcConference Proceedingsconference-collections
research-article

Fine-grained sentiment Feature Extraction Method for Cross-modal Sentiment Analysis

Published: 07 June 2024 Publication History

Abstract

It has been found that when people use images to express emotions, emotional information is often only strongly associated with some regions in the image, and these regions are also expressed in the corresponding language in people's comments. Therefore, the relationship between the emotional regions of an image and the relevant text is of great significance for cross-modal sentiment analysis. However, the existing methods simply use the object detection model to extract multiple object regions in the image, without effective screening, or the screening method is too coarse to introduce too much noise. Based on this observation, this paper proposes a novel image-text interactive filtering mechanism to capture the fine-grained features of sentiment, which is used for the screening of fine-grained sentiment regions in cross-modal sentiment analysis. Then, a sentiment consistency learning method is designed to obtain better sentiment feature encoding, so that the model has stronger sentiment classification ability. In addition, considering that the emotional regions extracted by object detection may not necessarily represent complete emotional information, we integrate the contextual feature representation of each individual modality to achieve more reliable prediction. In this paper, we name all the proposed methods Fine-Grained Sentiment Consistency Interaction Network (FSCIN) and achieve good performance improvement on three cross-modal sentiment analysis datasets, which proves the effectiveness of our method.

References

[1]
Zhu T, Li L, Yang J, Multimodal sentiment analysis with image-text interaction network[J]. IEEE Transactions on Multimedia, 2022.
[2]
Yu J, Jiang J. Adapting BERT for target-oriented multimodal sentiment classification[C]. IJCAI, 2019.
[3]
Kim W, Son B, Kim I. Vilt: Vision-and-language transformer without convolution or region supervision[C]//International Conference on Machine Learning. PMLR, 2021: 5583-5594.
[4]
Khan Z, Fu Y. Exploiting BERT for multimodal target sentiment classification through input space translation[C]//Proceedings of the 29th ACM International Conference on Multimedia. 2021: 3034-3042.
[5]
Xiao X, Pu Y, Zhou D, Collaborative fine-grained interaction learning for image–text sentiment analysis[J]. Knowledge-Based Systems, 2023, 279: 110951.
[6]
Zhu T, Li L, Yang J, Multimodal sentiment analysis with image-text interaction network[J]. IEEE Transactions on Multimedia, 2022.
[7]
Khan Z, Fu Y. Exploiting BERT for multimodal target sentiment classification through input space translation[C]//Proceedings of the 29th ACM international conference on multimedia. 2021: 3034-3042.
[8]
He K, Zhang X, Ren S, Deep residual learning for image recognition[C]//Proceedings of the IEEE conference on computer vision and pattern recognition. 2016: 770-778.
[9]
Dosovitskiy A, Beyer L, Kolesnikov A, An image is worth 16x16 words: Transformers for image recognition at scale[J]. arXiv preprint arXiv:2010.11929, 2020.
[10]
Kim Y. Convolutional neural networks for sentence classification[J]. arXiv preprint arXiv:1408.5882, 2014.
[11]
Lai S, Xu L, Liu K, Recurrent convolutional neural networks for text classification[C]//Proceedings of the AAAI conference on artificial intelligence. 2015, 29(1).
[12]
Yue L, Chen W, Li X, A survey of sentiment analysis in social media[J]. Knowledge and Information Systems, 2019, 60: 617-663.
[13]
Barrett L F, Lindquist K A, Gendron M. Language as context for the perception of emotion[J]. Trends in cognitive sciences, 2007, 11(8)

Index Terms

  1. Fine-grained sentiment Feature Extraction Method for Cross-modal Sentiment Analysis

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    ICMLC '24: Proceedings of the 2024 16th International Conference on Machine Learning and Computing
    February 2024
    757 pages
    ISBN:9798400709234
    DOI:10.1145/3651671
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 07 June 2024

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Cross-modal sentiment analysis
    2. Feature optimization for sentiment regions
    3. Fine-grained sentiment features

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    ICMLC 2024

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 37
      Total Downloads
    • Downloads (Last 12 months)37
    • Downloads (Last 6 weeks)6
    Reflects downloads up to 25 Dec 2024

    Other Metrics

    Citations

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media