March 2024
·
28 Reads
·
5 Citations
This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.
March 2024
·
28 Reads
·
5 Citations
August 2023
·
1,537 Reads
·
1 Citation
In our modern, fast-paced, and interconnected world, the importance of mental well-being has grown into a matter of great urgency. However, traditional methods such as Emotional Support Conversations (ESC) face challenges in effectively addressing a diverse range of individual personalities. In response, we introduce the Social Support Conversation (S2Conv) framework. It comprises a series of support agents and the interpersonal matching mechanism, linking individuals with persona-compatible virtual supporters. Utilizing persona decomposition based on the MBTI (Myers-Briggs Type Indicator), we have created the MBTI-1024 Bank, a group that of virtual characters with distinct profiles. Through improved role-playing prompts with behavior preset and dynamic memory, we facilitate the development of the MBTI-S2Conv dataset, which contains conversations between the characters in the MBTI-1024 Bank. Building upon these foundations, we present CharacterChat, a comprehensive S2Conv system, which includes a conversational model driven by personas and memories, along with an interpersonal matching plugin model that dispatches the optimal supporters from the MBTI-1024 Bank for individuals with specific personas. Empirical results indicate the remarkable efficacy of CharacterChat in providing personalized social support and highlight the substantial advantages derived from interpersonal matching. The source code is available in \url{https://github.com/morecry/CharacterChat}.
June 2023
·
73 Reads
·
26 Citations
Proceedings of the AAAI Conference on Artificial Intelligence
Previous works on emotion recognition in conversation (ERC) follow a two-step paradigm, which can be summarized as first producing context-independent features via fine-tuning pretrained language models (PLMs) and then analyzing contextual information and dialogue structure information among the extracted features. However, we discover that this paradigm has several limitations. Accordingly, we propose a novel paradigm, i.e., exploring contextual information and dialogue structure information in the fine-tuning step, and adapting the PLM to the ERC task in terms of input text, classification structure, and training strategy. Furthermore, we develop our model BERT-ERC according to the proposed paradigm, which improves ERC performance in three aspects, namely suggestive text, fine-grained classification module, and two-stage training. Compared to existing methods, BERT-ERC achieves substantial improvement on four datasets, indicating its effectiveness and generalization capability. Besides, we also set up the limited resources scenario and the online prediction scenario to approximate real-world scenarios. Extensive experiments demonstrate that the proposed paradigm significantly outperforms the previous one and can be adapted to various scenes.
June 2023
·
23 Reads
The standard definition generation task requires to automatically produce mono-lingual definitions (e.g., English definitions for English words), but ignores that the generated definitions may also consist of unfamiliar words for language learners. In this work, we propose a novel task of Trans-Lingual Definition Generation (TLDG), which aims to generate definitions in another language, i.e., the native speaker's language. Initially, we explore the unsupervised manner of this task and build up a simple implementation of fine-tuning the multi-lingual machine translation model. Then, we develop two novel methods, Prompt Combination and Contrastive Prompt Learning, for further enhancing the quality of the generation. Our methods are evaluated against the baseline Pipeline method in both rich- and low-resource settings, and we empirically establish its superiority in generating higher-quality trans-lingual definitions.
January 2023
·
42 Reads
Previous works on emotion recognition in conversation (ERC) follow a two-step paradigm, which can be summarized as first producing context-independent features via fine-tuning pretrained language models (PLMs) and then analyzing contextual information and dialogue structure information among the extracted features. However, we discover that this paradigm has several limitations. Accordingly, we propose a novel paradigm, i.e., exploring contextual information and dialogue structure information in the fine-tuning step, and adapting the PLM to the ERC task in terms of input text, classification structure, and training strategy. Furthermore, we develop our model BERT-ERC according to the proposed paradigm, which improves ERC performance in three aspects, namely suggestive text, fine-grained classification module, and two-stage training. Compared to existing methods, BERT-ERC achieves substantial improvement on four datasets, indicating its effectiveness and generalization capability. Besides, we also set up the limited resources scenario and the online prediction scenario to approximate real-world scenarios. Extensive experiments demonstrate that the proposed paradigm significantly outperforms the previous one and can be adapted to various scenes.
January 2023
·
7 Reads
·
8 Citations
January 2023
·
5 Reads
·
5 Citations
October 2022
·
7 Reads
Recently, pre-trained transformer-based models have achieved great success in the task of definition generation (DG). However, previous encoder-decoder models lack effective representation learning to contain full semantic components of the given word, which leads to generating under-specific definitions. To address this problem, we propose a novel contrastive learning method, encouraging the model to capture more detailed semantic representations from the definition sequence encoding. According to both automatic and manual evaluation, the experimental results on three mainstream benchmarks demonstrate that the proposed method could generate more specific and high-quality definitions compared with several state-of-the-art models.
July 2022
·
92 Reads
·
21 Citations
April 2022
·
69 Reads
Existing commonsense knowledge bases often organize tuples in an isolated manner, which is deficient for commonsense conversational models to plan the next steps. To fill the gap, we curate a large-scale multi-turn human-written conversation corpus, and create the first Chinese commonsense conversation knowledge graph which incorporates both social commonsense knowledge and dialog flow information. To show the potential of our graph, we develop a graph-conversation matching approach, and benchmark two graph-grounded conversational tasks.
... Previous works on definition generation mainly focus on mono-lingual generation scenarios, primarily due to the availability of parallel training and evaluation data (Yang et al., 2020;Zhang et al., 2022a). However, these works rarely notice a real-occurring problem that the generated definitions may also consist of unfamiliar words for language learners (Zhang, 2011). ...
January 2022
... Due to its significant practical applications, ERG has attracted substantial and sustained prior research attention [11,24,54]. Existing studies have developed various methods to enhance the performance of ERG systems [2,13,38,50]. ...
March 2024
... For example, Inoue et al. (2022) presented character understanding as a suite of document-level tasks that included gender and role identification of the character, cloze tasks, quote attribution and question answering. Li et al. (2023) adopted coreference resolution, character linking and speaker guessing tasks, and Azab et al. (2019) used character relationships and relatedness to evaluate character representations. We organized the character understanding tasks into the following categories. ...
January 2023
... However, this also poses substantial challenges to the automatic evaluation of these models. Representatively, LLM-based AI agents' focus transfer from traditional natural language processing tasks Zhang et al., 2023) to real-world (Liu et al., 2023b;Huang et al., 2023), open-ended response generation , which greatly limits the applicability of traditional n-gram matching methods (e.g., BLEU (Papineni et al., 2002) and ROUGE (Lin, 2004)) (Liu et al., 2016;Reiter, 2018) or model-based evaluators (Zhang et al., 2020;Zhong et al., 2022) for evaluation. ...
January 2023
... Additionally, advanced attention mechanisms, including multi-head self-attention and cross-attention, are employed to emphasize emotionally salient features and capture temporal emotional shifts effectively. To further enrich contextual embeddings, representations are transformed using BERT [24], GPT-3, and RoBERTa [25], and subsequently stored in a FAISS vector database [26], enabling the efficient retrieval of relevant information during dialogue generation. By integrating these techniques, our framework significantly enhances empathy, coherence, informativeness, and fluency in LLM-generated responses. ...
June 2023
Proceedings of the AAAI Conference on Artificial Intelligence
... Simile plays an important role in human language to make utterances more vivid, interesting, and graspable (Zhang et al., 2021; and is an increasingly studied phenomenon in computational linguistics (Song et al., 2021;. A simile is a figure of speech that compares two things from different categories (called the tenor and the vehicle) via shared properties (Paul, 1970). ...
May 2021
Proceedings of the AAAI Conference on Artificial Intelligence
... In recent years, conversational recommendation have gained growing attention within the research community of recommender systems, and various implementations of Conversational Recommender Systems (CRSs) have been proposed (e.g., [6,[27][28][29]32]). Unlike traditional Recommender Systems (RS), which model user interests and generate recommendations based on implicit feedback signals such as clicking or browsing history [1,9,10,19,20,25], CRSs solicit and identify user interests through natural language conversations to make recommendations [13,17,35]. ...
July 2022
... Some approaches inject commonsense knowledge to improve understanding of help-seekers (e.g. MISC (Tu et al. 2022), C3KG , GLHG (Peng et al. 2022)), while others employ cognitive reasoning to gradually infer the mental state of help-seekers (e.g. Dia-logueCoT (Chae et al. 2023), CueCoT (Wang et al. 2023)). ...
January 2022
... Using ATOMIC and Conceptnet, which are external knowledge graphs, authors capture different aspects of commonsense knowledge. Several studies [29], [30], [31] highlighted the use of the COMET model in different contexts, such as counseling, Chinese dialogue and emotional support. These models seek to enhance dialogue responses by integrating external commonsense knowledge, but exploiting commonsense knowledge for text generation is still relatively nascent. ...
January 2022
... Beyond the actual service provided, customers are more likely to trust an employee or organization when their emotional needs are met, particularly if these emotions are perceived as genuine. This understanding has motivated increased interest in embedding emotionalsensitivity into customer service chatbots (Bilquise et al., 2022;Li et al., 2021;Pamungkas, 2019). However, it is unclear if emotions from an AI would have the same effect. ...
July 2021