Leveraging Large Language Models for Predictive Analysis of Human Misery
Abstract
LLMs predict misery scores from text using various prompting strategies, with few-shot approaches outperforming zero-shot, and a gamified framework assessing their adaptability in emotional reasoning tasks.
This study investigates the use of Large Language Models (LLMs) for predicting human-perceived misery scores from natural language descriptions of real-world scenarios. The task is framed as a regression problem, where the model assigns a scalar value from 0 to 100 to each input statement. We evaluate multiple prompting strategies, including zero-shot, fixed-context few-shot, and retrieval-based prompting using BERT sentence embeddings. Few-shot approaches consistently outperform zero-shot baselines, underscoring the value of contextual examples in affective prediction. To move beyond static evaluation, we introduce the "Misery Game Show", a novel gamified framework inspired by a television format. It tests LLMs through structured rounds involving ordinal comparison, binary classification, scalar estimation, and feedback-driven reasoning. This setup enables us to assess not only predictive accuracy but also the model's ability to adapt based on corrective feedback. The gamified evaluation highlights the broader potential of LLMs in dynamic emotional reasoning tasks beyond standard regression. Code and data link: https://github.com/abhi1nandy2/Misery_Data_Exps_GitHub
Community
In our paper, we show how Large Language Models can predict misery scores from text. We also introduce the "Misery Game Show" - a fun, gamified test that not only checks accuracy but also how well LLMs adapt to feedback, opening new possibilities for emotional reasoning!
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Can LLMs Infer Personality from Real World Conversations? (2025)
- Large Language Models for Subjective Language Understanding: A Survey (2025)
- SpeechR: A Benchmark for Speech Reasoning in Large Audio-Language Models (2025)
- CAAD: Context-Aware Adaptive Decoding for Truthful Text Generation (2025)
- Can LLM-Generated Textual Explanations Enhance Model Classification Performance? An Empirical Study (2025)
- Commander-GPT: Dividing and Routing for Multimodal Sarcasm Detection (2025)
- AIGVE-MACS: Unified Multi-Aspect Commenting and Scoring Model for AI-Generated Video Evaluation (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper