Deprecated: The each() function is deprecated. This message will be suppressed on further calls in /home/zhenxiangba/zhenxiangba.com/public_html/phproxy-improved-master/index.php on line 456 Paper page - LongEmotion: Measuring Emotional Intelligence of Large Language Models
in Long-Context Interaction
Please give a thumbs up to this comment if you found it helpful!
\n
If you want recommendations for any Paper on Hugging Face checkout this Space
\n
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: \n\n@librarian-bot\n\t recommend
\n","updatedAt":"2025-09-17T01:35:20.763Z","author":{"_id":"63d3e0e8ff1384ce6c5dd17d","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1674830754237-63d3e0e8ff1384ce6c5dd17d.jpeg","fullname":"Librarian Bot (Bot)","name":"librarian-bot","type":"user","isPro":false,"isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":318,"isUserFollowing":false}},"numEdits":0,"identifiedLanguage":{"language":"en","probability":0.6971155405044556},"editors":["librarian-bot"],"editorAvatarUrls":["https://cdn-avatars.huggingface.co/v1/production/uploads/1674830754237-63d3e0e8ff1384ce6c5dd17d.jpeg"],"reactions":[],"isReport":false}}],"primaryEmailConfirmed":false,"paper":{"id":"2509.07403","authors":[{"_id":"68c398f9fc1747b912403aa3","user":{"_id":"66b09d0a7cb12daaffcba47e","avatarUrl":"/avatars/ed25d0a3d816dff4e86e7c59989f7277.svg","isPro":false,"fullname":"Matcha","user":"MatchaLwc","type":"user"},"name":"Weichu Liu","status":"claimed_verified","statusLastChangedAt":"2025-09-18T13:27:53.079Z","hidden":false},{"_id":"68c398f9fc1747b912403aa4","user":{"_id":"60851545a5da133ac6c38686","avatarUrl":"/avatars/d385fcc513acef80a3b711aa92d898e5.svg","isPro":false,"fullname":"Jing Xiong","user":"menik1126","type":"user"},"name":"Jing Xiong","status":"claimed_verified","statusLastChangedAt":"2025-09-12T16:07:35.882Z","hidden":false},{"_id":"68c398f9fc1747b912403aa5","name":"Yuxuan Hu","hidden":false},{"_id":"68c398f9fc1747b912403aa6","name":"Zixuan Li","hidden":false},{"_id":"68c398f9fc1747b912403aa7","name":"Minghuan Tan","hidden":false},{"_id":"68c398f9fc1747b912403aa8","name":"Ningning Mao","hidden":false},{"_id":"68c398f9fc1747b912403aa9","name":"Chenyang Zhao","hidden":false},{"_id":"68c398f9fc1747b912403aaa","name":"Zhongwei Wan","hidden":false},{"_id":"68c398f9fc1747b912403aab","user":{"_id":"60fc2fcca6bdebbe52dfdaf4","avatarUrl":"/avatars/1d59a7f33cb0df04678516f337e6b881.svg","isPro":false,"fullname":"Chaofan Tao","user":"tcftrees","type":"user"},"name":"Chaofan Tao","status":"claimed_verified","statusLastChangedAt":"2025-11-24T12:27:13.468Z","hidden":false},{"_id":"68c398f9fc1747b912403aac","name":"Wendong Xu","hidden":false},{"_id":"68c398f9fc1747b912403aad","user":{"_id":"640feb91b0ee289c8583dd59","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/640feb91b0ee289c8583dd59/CA4UZj2mEsJO2UR0NnN5H.jpeg","isPro":false,"fullname":"Hui Shen","user":"Cloudriver","type":"user"},"name":"Hui Shen","status":"claimed_verified","statusLastChangedAt":"2025-09-15T15:08:38.540Z","hidden":false},{"_id":"68c398f9fc1747b912403aae","name":"Chengming Li","hidden":false},{"_id":"68c398f9fc1747b912403aaf","name":"Lingpeng Kong","hidden":false},{"_id":"68c398f9fc1747b912403ab0","name":"Ngai Wong","hidden":false}],"publishedAt":"2025-09-09T05:32:45.000Z","submittedOnDailyAt":"2025-09-16T17:53:39.128Z","title":"LongEmotion: Measuring Emotional Intelligence of Large Language Models\n in Long-Context Interaction","submittedOnDailyBy":{"_id":"640feb91b0ee289c8583dd59","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/640feb91b0ee289c8583dd59/CA4UZj2mEsJO2UR0NnN5H.jpeg","isPro":false,"fullname":"Hui Shen","user":"Cloudriver","type":"user"},"summary":"Large language models (LLMs) make significant progress in Emotional\nIntelligence (EI) and long-context understanding. However, existing benchmarks\ntend to overlook certain aspects of EI in long-context scenarios, especially\nunder realistic, practical settings where interactions are lengthy, diverse,\nand often noisy. To move towards such realistic settings, we present\nLongEmotion, a benchmark specifically designed for long-context EI tasks. It\ncovers a diverse set of tasks, including Emotion Classification, Emotion\nDetection, Emotion QA, Emotion Conversation, Emotion Summary, and Emotion\nExpression. On average, the input length for these tasks reaches 8,777 tokens,\nwith long-form generation required for Emotion Expression. To enhance\nperformance under realistic constraints, we incorporate Retrieval-Augmented\nGeneration (RAG) and Collaborative Emotional Modeling (CoEM), and compare them\nwith standard prompt-based methods. Unlike conventional approaches, our RAG\nmethod leverages both the conversation context and the large language model\nitself as retrieval sources, avoiding reliance on external knowledge bases. The\nCoEM method further improves performance by decomposing the task into five\nstages, integrating both retrieval augmentation and limited knowledge\ninjection. Experimental results show that both RAG and CoEM consistently\nenhance EI-related performance across most long-context tasks, advancing LLMs\ntoward more practical and real-world EI applications. Furthermore, we conducted\na comparative case study experiment on the GPT series to demonstrate the\ndifferences among various models in terms of EI. Code is available on GitHub at\nhttps://github.com/LongEmotion/LongEmotion, and the project page can be found\nat https://longemotion.github.io/.","upvotes":35,"discussionId":"68c398f9fc1747b912403ab1","projectPage":"https://longemotion.github.io/","githubRepo":"https://github.com/LongEmotion/LongEmotion","githubRepoAddedBy":"user","ai_summary":"LongEmotion benchmark enhances large language models' emotional intelligence in long-context scenarios using Retrieval-Augmented Generation and Collaborative Emotional Modeling.","ai_keywords":["Large language models","Emotional Intelligence","long-context understanding","LongEmotion","Emotion Classification","Emotion Detection","Emotion QA","Emotion Conversation","Emotion Summary","Emotion Expression","Retrieval-Augmented Generation","Collaborative Emotional Modeling","GPT series"],"githubStars":9,"organization":{"_id":"68aff89c4016e55f8903cfce","name":"longemotion1","fullname":"LongEmotion","avatar":"https://cdn-uploads.huggingface.co/production/uploads/68afeb0c955b4d99967703d1/AbQqBfIuWLJlRl5hRlDXu.png"}},"canReadDatabase":false,"canManagePapers":false,"canSubmit":false,"hasHfLevelAccess":false,"upvoted":false,"upvoters":[{"_id":"640feb91b0ee289c8583dd59","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/640feb91b0ee289c8583dd59/CA4UZj2mEsJO2UR0NnN5H.jpeg","isPro":false,"fullname":"Hui Shen","user":"Cloudriver","type":"user"},{"_id":"60851545a5da133ac6c38686","avatarUrl":"/avatars/d385fcc513acef80a3b711aa92d898e5.svg","isPro":false,"fullname":"Jing Xiong","user":"menik1126","type":"user"},{"_id":"65437ef98f84f869869cb698","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/65437ef98f84f869869cb698/xRFcyXwQq7szB-GfAeoHu.png","isPro":false,"fullname":"Wendong XU","user":"kirailol","type":"user"},{"_id":"65df33cc172353c169dceaaf","avatarUrl":"/avatars/938f4fcbea0305279ca9ec37ce7eaa65.svg","isPro":false,"fullname":"Ubec","user":"hrw","type":"user"},{"_id":"68afeb0c955b4d99967703d1","avatarUrl":"/avatars/e366a0b1768e46bbc568006fbefa63be.svg","isPro":false,"fullname":"LongEmotion","user":"LongEmotion","type":"user"},{"_id":"66b09d0a7cb12daaffcba47e","avatarUrl":"/avatars/ed25d0a3d816dff4e86e7c59989f7277.svg","isPro":false,"fullname":"Matcha","user":"MatchaLwc","type":"user"},{"_id":"64955274003c1f54c97bf195","avatarUrl":"/avatars/e442cec1f25ad934e50d6fa3bb5572c1.svg","isPro":false,"fullname":"yukang lin","user":"linyk","type":"user"},{"_id":"65fbd50b04769daf218c07bc","avatarUrl":"/avatars/3c7a0059f76ea4cc58ce3c86c55b5414.svg","isPro":false,"fullname":"Han","user":"Tanrn","type":"user"},{"_id":"65950d64a02954c982f1938c","avatarUrl":"/avatars/ddbdb2b1bafbbd91f19ab3ba247c2152.svg","isPro":false,"fullname":"Yuxuan Hu","user":"huyx55","type":"user"},{"_id":"68cb9df34f4a1865541019ea","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/no-auth/PCy05ApSPfdgAOe1rPbT_.png","isPro":false,"fullname":"原嘉潞","user":"stitchlove2025","type":"user"},{"_id":"60fc2fcca6bdebbe52dfdaf4","avatarUrl":"/avatars/1d59a7f33cb0df04678516f337e6b881.svg","isPro":false,"fullname":"Chaofan Tao","user":"tcftrees","type":"user"},{"_id":"6305686a80bc5e03dad2e5b6","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/1661298779154-noauth.jpeg","isPro":false,"fullname":"zhaochenyang20","user":"zhaochenyang20","type":"user"}],"acceptLanguages":["*"],"dailyPaperRank":2,"organization":{"_id":"68aff89c4016e55f8903cfce","name":"longemotion1","fullname":"LongEmotion","avatar":"https://cdn-uploads.huggingface.co/production/uploads/68afeb0c955b4d99967703d1/AbQqBfIuWLJlRl5hRlDXu.png"}}">
LongEmotion benchmark enhances large language models' emotional intelligence in long-context scenarios using Retrieval-Augmented Generation and Collaborative Emotional Modeling.
AI-generated summary
Large language models (LLMs) make significant progress in Emotional
Intelligence (EI) and long-context understanding. However, existing benchmarks
tend to overlook certain aspects of EI in long-context scenarios, especially
under realistic, practical settings where interactions are lengthy, diverse,
and often noisy. To move towards such realistic settings, we present
LongEmotion, a benchmark specifically designed for long-context EI tasks. It
covers a diverse set of tasks, including Emotion Classification, Emotion
Detection, Emotion QA, Emotion Conversation, Emotion Summary, and Emotion
Expression. On average, the input length for these tasks reaches 8,777 tokens,
with long-form generation required for Emotion Expression. To enhance
performance under realistic constraints, we incorporate Retrieval-Augmented
Generation (RAG) and Collaborative Emotional Modeling (CoEM), and compare them
with standard prompt-based methods. Unlike conventional approaches, our RAG
method leverages both the conversation context and the large language model
itself as retrieval sources, avoiding reliance on external knowledge bases. The
CoEM method further improves performance by decomposing the task into five
stages, integrating both retrieval augmentation and limited knowledge
injection. Experimental results show that both RAG and CoEM consistently
enhance EI-related performance across most long-context tasks, advancing LLMs
toward more practical and real-world EI applications. Furthermore, we conducted
a comparative case study experiment on the GPT series to demonstrate the
differences among various models in terms of EI. Code is available on GitHub at
https://github.com/LongEmotion/LongEmotion, and the project page can be found
at https://longemotion.github.io/.
LongEmotion benchmark enhances large language models' emotional intelligence in long-context scenarios using Retrieval-Augmented Generation and Collaborative Emotional Modeling.