Skip to content

Emo Pillars tools for sentiment analysis NEW

+ WHO?

Universitat Pompeu Fabra (UPF)
Partner

+ WHAT?

Sentiment analysis from crawled social media data
and reviews

+ HOW?

Fine-grained context-less and contextual emotion classification using transformer-based neural networks trained on synthetic data generated by large language models

Concept

It is of high importance for creators of novel acoustic experiences to receive feedback from their audiences, including those in a virtual setting. Given user comments about the work on social media, it is desirable to automatically get a detailed view of people’s perceptions, including their explicitly expressed feelings. It is crucial to identify emotions within a wide spectrum of emotional categories, so as not to miss essential details that were communicated through language. As comments on social media are often brief, it is also important to properly situate them in context to avoid ambiguous emotional interpretations. To address this need in ReSilence, we have developed an Emo Pillars suite – a range of models for fine-grained emotion classification (with and without context), and released them along with our synthetic dataset, on which they were trained. We expect that our Emo Pillars models will be of high interest to various stakeholders in the concert creation industry, as well as to the authors of artistic implementations.

Tools

Emo Pillars is a collection of neural multi-label classifiers for 28 emotional classes. The project includes:

Emo Pillars Models

The Emo Pillars collection contains several specialized models, with usage and evaluation details available on their respective pages:

  • Context-Aware Models:
    • 28 Emotional Classes: A multi-label classifier that takes context and a character description to extract emotions from an utterance.
    • Fine-Tuned on EmoContext: A model fine-tuned on a 4-class EmoContext dataset (angry, sad, happy, and others). It can analyze either a context-plus-utterance or a three-turn dialogue.
  • Context-Less Models:
Experiments

Our Emo Pillars models, dataset, and pipeline that generates diverse, labelled synthetic data by extracting knowledge from LLMs for fine-grained context-less and context-aware emotion classification were presented at the 63rd Annual Meeting of the Association for Computational Linguistics in July 2025 (ACL 2025, Vienna, Austria): https://2025.aclweb.org/

Lessons learned:
Main results and evaluations can be found in our paper in Findings of the Association for Computational Linguistics (ACL 2025): https://aclanthology.org/2025.findings-acl.10/

Benefits

Most datasets for sentiment analysis lack the context in which an opinion was expressed, which is often crucial for understanding emotions, and are mainly limited by a few emotion categories. Foundation large language models suffer from over-predicting emotions and are too resource-intensive. When used for synthetic data generation, the produced examples generally lack semantic diversity.

Our Emo Pillars models are accessible, lightweight BERT-type encoder models trained on data from our LLM-based data synthesis pipeline, which focuses on expanding the semantic diversity of examples. The pipeline grounds emotional text generation in a corpus of narratives, resulting in non-repetitive utterances with unique contexts across 28 emotion classes.

Emo Pillars also include task-specific models, which are derived from training our base models on specific downstream tasks. The evaluation scores demonstrate that our base models are highly adaptable to new domains.

Along with the models, we release a dataset of 100,000 contextual and 300,000 contextless examples for fine-tuning various types of pre-trained language models. We validated our dataset through statistical analysis and human evaluation, confirming the success of our measures in diversifying utterances and contexts.

Adaptability / interoperability:
The outcomes of our models, which include sentiments with their expressiveness levels, can be easily integrated into personalized report generation systems to summarize emotion-aspect associations represented in a collection of user comments and produce structured analysis with insights and suggestions for specific roles organizing artistic performances.

Links

Hugging face

ACL anthology

Video
Project