Saturday, May 16, 2026
Science
No Result
View All Result
  • Login
  • HOME
  • SCIENCE NEWS
  • CONTACT US
  • HOME
  • SCIENCE NEWS
  • CONTACT US
No Result
View All Result
Scienmag
No Result
View All Result
Home Science News Social Science

New AI framework enhances emotion analysis

June 26, 2024
in Social Science
Reading Time: 3 mins read
0
Overall architecture of the two-stage stacked transformer framework
67
SHARES
608
VIEWS
Share on FacebookShare on Twitter
ADVERTISEMENT

Social media enthusiasts tend to spice up their text posts with emojis, images, audio, or video to attract more attention. Simple as it is, this technique makes scientific sense: multimodal information is found to be more effective in conveying emotions, as different modalities interact and enhance one another. To advance the understanding of these interactions and improve the analysis of emotions expressed through modality combinations, a Chinese research team introduced a novel two-stage framework using two stacked layers of transformers, state-of-the-art AI models for multimodal sentiment analysis. This study was published May 24 in Intelligent Computing, a Science Partner Journal.

Overall architecture of the two-stage stacked transformer framework

Credit: GUOFENG YI ET AL.

Social media enthusiasts tend to spice up their text posts with emojis, images, audio, or video to attract more attention. Simple as it is, this technique makes scientific sense: multimodal information is found to be more effective in conveying emotions, as different modalities interact and enhance one another. To advance the understanding of these interactions and improve the analysis of emotions expressed through modality combinations, a Chinese research team introduced a novel two-stage framework using two stacked layers of transformers, state-of-the-art AI models for multimodal sentiment analysis. This study was published May 24 in Intelligent Computing, a Science Partner Journal.

Current research in multimodal sentiment analysis often focuses on either fusing different modalities or addressing intricate interactions or adaptations between different types of fused information. Either approach alone can lead to information loss. This team’s framework, on the other hand, fuses information in two stages to effectively capture information on both levels. It was tested on three open datasets—MOSI, MOSEI, and SIMS—and performed better than or as well as the benchmark models.

The general workflow of this framework includes feature extraction, two stages of information fusion, and emotion prediction. First, text, audio, and video signals taken from source video clips are processed through their corresponding feature extractors and then encoded with additional context information into context-aware representations. Next, the three types of representations fuse for the first time: the text representations interact with the audio and video representations, allowing each modality to adapt to the others during the process, and the results further integrate with the original text representations. The text-centered output from the first stage then fuses with the adapted non-text representations so that they can enhance each other before the final, enriched output is ready for the emotion prediction stage.

The core of the team’s framework is stacked transformers and consists of bidirectional cross-modal transformers and a transformer encoder. These components correspond to two functional layers: the bidirectional interaction layer allows cross-modal interaction and is where the first-stage fusion occurs, and the refine layer addresses the more nuanced second-stage fusion.

To enhance the performance of the framework, the team implemented an attention weight accumulation mechanism that aggregates the attention weights of the text and non-text modalities during fusion to extract deeper shared information. Attention, a key concept in transformers, enables the model to identify and focus on the most relevant parts of the data. The team’s stacked transformers adopt two types of attention mechanism: the bidirectional cross-modal transformers use cross-attention, and the transformer encoder uses self-attention.

The future work of the team will focus on integrating more advanced transformers to improve computational efficiency and mitigate the inherent challenges associated with the self-attention mechanism.



Journal

Intelligent Computing

DOI

10.34133/icomputing.0081

Article Title

A Two-Stage Stacked Transformer Framework for Multimodal Sentiment Analysis

Article Publication Date

24-May-2024

Share27Tweet17
Previous Post

A chip-scale Titanium-sapphire laser

Next Post

Targeting human coatomer subunit zeta-1 by a marine natural product derivative represents a new therapeutic strategy to inhibit enterovirus A71 replication

Related Posts

Survey Uncovers How Biodiversity Education Shapes Classroom Learning — Social Science
Social Science

Survey Uncovers How Biodiversity Education Shapes Classroom Learning

May 16, 2026
Alexithymia in Schizophrenia and Dissociative Disorders — Social Science
Social Science

Alexithymia in Schizophrenia and Dissociative Disorders

May 16, 2026
Inflammation and Vascular Changes Linked to Early Schizophrenia — Social Science
Social Science

Inflammation and Vascular Changes Linked to Early Schizophrenia

May 15, 2026
Fear Learning’s Impact on Psychosis: EEG Study Insights — Social Science
Social Science

Fear Learning’s Impact on Psychosis: EEG Study Insights

May 15, 2026
Unraveling Worldwide Views on AI and Leadership in Education — Social Science
Social Science

Unraveling Worldwide Views on AI and Leadership in Education

May 15, 2026
Climate Change’s Impact on Social Health Explored — Social Science
Social Science

Climate Change’s Impact on Social Health Explored

May 15, 2026
Next Post
Chemoproteomics enables identification of coatomer subunit zeta-1 targeted by a small molecule for enterovirus A71 inhibition

Targeting human coatomer subunit zeta-1 by a marine natural product derivative represents a new therapeutic strategy to inhibit enterovirus A71 replication

  • Mothers who receive childcare support from maternal grandparents show more parental warmth, finds NTU Singapore study

    Mothers who receive childcare support from maternal grandparents show more parental warmth, finds NTU Singapore study

    27644 shares
    Share 11054 Tweet 6909
  • University of Seville Breaks 120-Year-Old Mystery, Revises a Key Einstein Concept

    1048 shares
    Share 419 Tweet 262
  • Bee body mass, pathogens and local climate influence heat tolerance

    678 shares
    Share 271 Tweet 170
  • Researchers record first-ever images and data of a shark experiencing a boat strike

    542 shares
    Share 217 Tweet 136
  • Groundbreaking Clinical Trial Reveals Lubiprostone Enhances Kidney Function

    528 shares
    Share 211 Tweet 132
Science

Embark on a thrilling journey of discovery with Scienmag.com—your ultimate source for cutting-edge breakthroughs. Immerse yourself in a world where curiosity knows no limits and tomorrow’s possibilities become today’s reality!

RECENT NEWS

  • Africa’s Regional Rainfall Extremes Set to Surge
  • Mobile App Enhances Exercise for Older Adults’ Cognition
  • Survey Uncovers How Biodiversity Education Shapes Classroom Learning
  • New Actinomycin Derivatives Target Malaria More Selectively

Categories

  • Agriculture
  • Anthropology
  • Archaeology
  • Athmospheric
  • Biology
  • Biotechnology
  • Blog
  • Bussines
  • Cancer
  • Chemistry
  • Climate
  • Earth Science
  • Editorial Policy
  • Marine
  • Mathematics
  • Medicine
  • Pediatry
  • Policy
  • Psychology & Psychiatry
  • Science Education
  • Social Science
  • Space
  • Technology and Engineering

Subscribe to Blog via Email

Enter your email address to subscribe to this blog and receive notifications of new posts by email.

Join 5,146 other subscribers

© 2025 Scienmag - Science Magazine

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In
No Result
View All Result
  • HOME
  • SCIENCE NEWS
  • CONTACT US

© 2025 Scienmag - Science Magazine

Discover more from Science

Subscribe now to keep reading and get access to the full archive.

Continue reading