Multi-views Emotional Knowledge Extraction for Emotion Recognition in Conversation
| dc.contributor.author | Jian Z. | |
| dc.contributor.author | Wu D. | |
| dc.contributor.author | Wang S. | |
| dc.contributor.author | He J. | |
| dc.contributor.author | Yao J. | |
| dc.contributor.author | Liu K. | |
| dc.contributor.author | Wu Q. | |
| dc.contributor.correspondence | Jian Z. | |
| dc.contributor.other | Mahidol University | |
| dc.date.accessioned | 2025-05-26T18:12:33Z | |
| dc.date.available | 2025-05-26T18:12:33Z | |
| dc.date.issued | 2025-07-08 | |
| dc.description.abstract | Emotion Recognition in Conversation (ERC) is a challenging task due to the scarcity and dispersion of contextual information across utterances. Most existing methods attempt to integrate comprehensive information to enhance utterance semantics, which, however, also introduces noise and irrelevant content, misleading the model and limiting its potential in emotion recognition. To this end, we introduce the concept of Conversational Clique (ConvClique) and propose CC-ERC, a multi-view emotional knowledge extraction method designed to capture the most relevant emotional cues within the ConvClique from complementary perspectives and collaboratively predict utterance emotions. Specifically, CC-ERC comprises two modules: 1) the Utterance Spatial Relationship (USR) module, which predicts emotions by modeling structural correlations among utterances, and 2) the Emotion Temporal Relationship (ETR) module, which captures emotion sequence patterns to determine utterance emotions. These modules are integrated to obtain the final prediction, contributing to the robustness and accuracy of emotion recognition. The effectiveness of CC-ERC is validated on three widely used ERC datasets, evaluated in both online and offline settings. Compared to the state-of-the-art methods, CC-ERC achieves average improvements of 0.63% in accuracy and 0.94% in weighted F1 scores. Ablation studies further validate the significance of ConvClique-based knowledge extraction and demonstrate the effectiveness of the USR and ETR modules in modeling utterance structural correlations and emotion sequence patterns. | |
| dc.identifier.citation | Knowledge-Based Systems Vol.322 (2025) | |
| dc.identifier.doi | 10.1016/j.knosys.2025.113601 | |
| dc.identifier.issn | 09507051 | |
| dc.identifier.scopus | 2-s2.0-105005491351 | |
| dc.identifier.uri | https://repository.li.mahidol.ac.th/handle/123456789/110377 | |
| dc.rights.holder | SCOPUS | |
| dc.subject | Business, Management and Accounting | |
| dc.subject | Computer Science | |
| dc.subject | Decision Sciences | |
| dc.title | Multi-views Emotional Knowledge Extraction for Emotion Recognition in Conversation | |
| dc.type | Article | |
| mu.datasource.scopus | https://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=105005491351&origin=inward | |
| oaire.citation.title | Knowledge-Based Systems | |
| oaire.citation.volume | 322 | |
| oairecerif.author.affiliation | College of Management Mahidol University | |
| oairecerif.author.affiliation | Xiamen University |
