Yanran Li’s research while affiliated with ‎Beijing Xiaomi Technology and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (20)


Empathetic Response Generation with Relation-aware Commonsense Knowledge
  • Conference Paper

March 2024

·

28 Reads

·

5 Citations

Changyu Chen

·

Yanran Li

·

·

[...]

·


Figure 2: Persona Decomposition Prompt Based on MBTI.
Figure 4: The Overview of Prompting ChatGPT for Engagement in Social Support Conversations through Role-Playing Prompts with the Behavior Preset and Dynamic Memory. (1) We randomly sample two characters from the MBTI-1024 Bank to act as the seeker and the supporter respectively. Then we transform their structured personas into role-playing prompts. (2) We introduce a behavior preset method to pre-define possible single-turn dialogue demonstrations, enabling ChatGPT to maintain the character's state for multi-turn conversations. (3) The context-related aspect of memory will be dynamically chosen as the reference for response generation in each turn. (4) The ChatGPT agent will generate responses based on the persona prompt and the dynamic selected memory.
Figure 5: Expiration Ratio of Role-Playing Prompts with Increasing Conversation Turns.
Figure 7: The description for 16 MBTI personalities.
CharacterChat: Learning towards Conversational AI with Personalized Social Support
  • Preprint
  • File available

August 2023

·

1,537 Reads

·

1 Citation

In our modern, fast-paced, and interconnected world, the importance of mental well-being has grown into a matter of great urgency. However, traditional methods such as Emotional Support Conversations (ESC) face challenges in effectively addressing a diverse range of individual personalities. In response, we introduce the Social Support Conversation (S2Conv) framework. It comprises a series of support agents and the interpersonal matching mechanism, linking individuals with persona-compatible virtual supporters. Utilizing persona decomposition based on the MBTI (Myers-Briggs Type Indicator), we have created the MBTI-1024 Bank, a group that of virtual characters with distinct profiles. Through improved role-playing prompts with behavior preset and dynamic memory, we facilitate the development of the MBTI-S2Conv dataset, which contains conversations between the characters in the MBTI-1024 Bank. Building upon these foundations, we present CharacterChat, a comprehensive S2Conv system, which includes a conversational model driven by personas and memories, along with an interpersonal matching plugin model that dispatches the optimal supporters from the MBTI-1024 Bank for individuals with specific personas. Empirical results indicate the remarkable efficacy of CharacterChat in providing personalized social support and highlight the substantial advantages derived from interpersonal matching. The source code is available in \url{https://github.com/morecry/CharacterChat}.

Download

BERT-ERC: Fine-Tuning BERT Is Enough for Emotion Recognition in Conversation

June 2023

·

73 Reads

·

26 Citations

Proceedings of the AAAI Conference on Artificial Intelligence

Previous works on emotion recognition in conversation (ERC) follow a two-step paradigm, which can be summarized as first producing context-independent features via fine-tuning pretrained language models (PLMs) and then analyzing contextual information and dialogue structure information among the extracted features. However, we discover that this paradigm has several limitations. Accordingly, we propose a novel paradigm, i.e., exploring contextual information and dialogue structure information in the fine-tuning step, and adapting the PLM to the ERC task in terms of input text, classification structure, and training strategy. Furthermore, we develop our model BERT-ERC according to the proposed paradigm, which improves ERC performance in three aspects, namely suggestive text, fine-grained classification module, and two-stage training. Compared to existing methods, BERT-ERC achieves substantial improvement on four datasets, indicating its effectiveness and generalization capability. Besides, we also set up the limited resources scenario and the online prediction scenario to approximate real-world scenarios. Extensive experiments demonstrate that the proposed paradigm significantly outperforms the previous one and can be adapted to various scenes.


Figure 1: The application scenes of a Chinese native speaker learning English and English native speaker learning Chinese. We also build a Chrome extension (in Appendix A) to better show the application scenes.
Figure 2: The area surrounded by the red dotted line represents the training process and the green dash line represents the inference process. In the training phase, (1) the task prompt will mix the language information from the language prompt (in blended color), and (2) the contrastive loss (upper right corner) is together applied with the MLE loss (upper left corner) to train the model jointly. At the inference stage, the language prompt could be set to any other languages used in the training stage for trans-lingual definition generation. Best viewed in color.
Figure 3: The application scene of Learning English words based on our best method. Given the word "struggles" and press the shortcut key, the application will identify its corresponding context and output the definition "表示给 某人带来困难。" (To make someone difficult.).
Generated result comparison between four methods on M2M model.
Assisting Language Learners: Automated Trans-Lingual Definition Generation via Contrastive Prompt Learning

June 2023

·

23 Reads

The standard definition generation task requires to automatically produce mono-lingual definitions (e.g., English definitions for English words), but ignores that the generated definitions may also consist of unfamiliar words for language learners. In this work, we propose a novel task of Trans-Lingual Definition Generation (TLDG), which aims to generate definitions in another language, i.e., the native speaker's language. Initially, we explore the unsupervised manner of this task and build up a simple implementation of fine-tuning the multi-lingual machine translation model. Then, we develop two novel methods, Prompt Combination and Contrastive Prompt Learning, for further enhancing the quality of the generation. Our methods are evaluated against the baseline Pipeline method in both rich- and low-resource settings, and we empirically establish its superiority in generating higher-quality trans-lingual definitions.


BERT-ERC: Fine-tuning BERT is Enough for Emotion Recognition in Conversation

January 2023

·

42 Reads

Previous works on emotion recognition in conversation (ERC) follow a two-step paradigm, which can be summarized as first producing context-independent features via fine-tuning pretrained language models (PLMs) and then analyzing contextual information and dialogue structure information among the extracted features. However, we discover that this paradigm has several limitations. Accordingly, we propose a novel paradigm, i.e., exploring contextual information and dialogue structure information in the fine-tuning step, and adapting the PLM to the ERC task in terms of input text, classification structure, and training strategy. Furthermore, we develop our model BERT-ERC according to the proposed paradigm, which improves ERC performance in three aspects, namely suggestive text, fine-grained classification module, and two-stage training. Compared to existing methods, BERT-ERC achieves substantial improvement on four datasets, indicating its effectiveness and generalization capability. Besides, we also set up the limited resources scenario and the online prediction scenario to approximate real-world scenarios. Extensive experiments demonstrate that the proposed paradigm significantly outperforms the previous one and can be adapted to various scenes.




Fine-grained Contrastive Learning for Definition Generation

October 2022

·

7 Reads

Recently, pre-trained transformer-based models have achieved great success in the task of definition generation (DG). However, previous encoder-decoder models lack effective representation learning to contain full semantic components of the given word, which leads to generating under-specific definitions. To address this problem, we propose a novel contrastive learning method, encouraging the model to capture more detailed semantic representations from the definition sequence encoding. According to both automatic and manual evaluation, the experimental results on three mainstream benchmarks demonstrate that the proposed method could generate more specific and high-quality definitions compared with several state-of-the-art models.



C3KG: A Chinese Commonsense Conversation Knowledge Graph

April 2022

·

69 Reads

Existing commonsense knowledge bases often organize tuples in an isolated manner, which is deficient for commonsense conversational models to plan the next steps. To fill the gap, we curate a large-scale multi-turn human-written conversation corpus, and create the first Chinese commonsense conversation knowledge graph which incorporates both social commonsense knowledge and dialog flow information. To show the potential of our graph, we develop a graph-conversation matching approach, and benchmark two graph-grounded conversational tasks.


Citations (11)


... Previous works on definition generation mainly focus on mono-lingual generation scenarios, primarily due to the availability of parallel training and evaluation data (Yang et al., 2020;Zhang et al., 2022a). However, these works rarely notice a real-occurring problem that the generated definitions may also consist of unfamiliar words for language learners (Zhang, 2011). ...

Reference:

Assisting Language Learners: Automated Trans-Lingual Definition Generation via Contrastive Prompt Learning
Fine-grained Contrastive Learning for Definition Generation
  • Citing Conference Paper
  • January 2022

... Due to its significant practical applications, ERG has attracted substantial and sustained prior research attention [11,24,54]. Existing studies have developed various methods to enhance the performance of ERG systems [2,13,38,50]. ...

Empathetic Response Generation with Relation-aware Commonsense Knowledge
  • Citing Conference Paper
  • March 2024

... For example, Inoue et al. (2022) presented character understanding as a suite of document-level tasks that included gender and role identification of the character, cloze tasks, quote attribution and question answering. Li et al. (2023) adopted coreference resolution, character linking and speaker guessing tasks, and Azab et al. (2019) used character relationships and relatedness to evaluate character representations. We organized the character understanding tasks into the following categories. ...

Multi-level Contrastive Learning for Script-based Character Understanding

... However, this also poses substantial challenges to the automatic evaluation of these models. Representatively, LLM-based AI agents' focus transfer from traditional natural language processing tasks Zhang et al., 2023) to real-world (Liu et al., 2023b;Huang et al., 2023), open-ended response generation , which greatly limits the applicability of traditional n-gram matching methods (e.g., BLEU (Papineni et al., 2002) and ROUGE (Lin, 2004)) (Liu et al., 2016;Reiter, 2018) or model-based evaluators (Zhang et al., 2020;Zhong et al., 2022) for evaluation. ...

Assisting Language Learners: Automated Trans-Lingual Definition Generation via Contrastive Prompt Learning

... Additionally, advanced attention mechanisms, including multi-head self-attention and cross-attention, are employed to emphasize emotionally salient features and capture temporal emotional shifts effectively. To further enrich contextual embeddings, representations are transformed using BERT [24], GPT-3, and RoBERTa [25], and subsequently stored in a FAISS vector database [26], enabling the efficient retrieval of relevant information during dialogue generation. By integrating these techniques, our framework significantly enhances empathy, coherence, informativeness, and fluency in LLM-generated responses. ...

BERT-ERC: Fine-Tuning BERT Is Enough for Emotion Recognition in Conversation
  • Citing Article
  • June 2023

Proceedings of the AAAI Conference on Artificial Intelligence

... Simile plays an important role in human language to make utterances more vivid, interesting, and graspable (Zhang et al., 2021; and is an increasingly studied phenomenon in computational linguistics (Song et al., 2021;. A simile is a figure of speech that compares two things from different categories (called the tenor and the vehicle) via shared properties (Paul, 1970). ...

Writing Polishment with Simile: Task, Dataset and A Neural Approach
  • Citing Article
  • May 2021

Proceedings of the AAAI Conference on Artificial Intelligence

... In recent years, conversational recommendation have gained growing attention within the research community of recommender systems, and various implementations of Conversational Recommender Systems (CRSs) have been proposed (e.g., [6,[27][28][29]32]). Unlike traditional Recommender Systems (RS), which model user interests and generate recommendations based on implicit feedback signals such as clicking or browsing history [1,9,10,19,20,25], CRSs solicit and identify user interests through natural language conversations to make recommendations [13,17,35]. ...

Conversational Recommendation via Hierarchical Information Modeling
  • Citing Conference Paper
  • July 2022

... Some approaches inject commonsense knowledge to improve understanding of help-seekers (e.g. MISC (Tu et al. 2022), C3KG , GLHG (Peng et al. 2022)), while others employ cognitive reasoning to gradually infer the mental state of help-seekers (e.g. Dia-logueCoT (Chae et al. 2023), CueCoT (Wang et al. 2023)). ...

MISC: A Mixed Strategy-Aware Model integrating COMET for Emotional Support Conversation
  • Citing Conference Paper
  • January 2022

... Using ATOMIC and Conceptnet, which are external knowledge graphs, authors capture different aspects of commonsense knowledge. Several studies [29], [30], [31] highlighted the use of the COMET model in different contexts, such as counseling, Chinese dialogue and emotional support. These models seek to enhance dialogue responses by integrating external commonsense knowledge, but exploiting commonsense knowledge for text generation is still relatively nascent. ...

C3KG: A Chinese Commonsense Conversation Knowledge Graph
  • Citing Conference Paper
  • January 2022

... Beyond the actual service provided, customers are more likely to trust an employee or organization when their emotional needs are met, particularly if these emotions are perceived as genuine. This understanding has motivated increased interest in embedding emotionalsensitivity into customer service chatbots (Bilquise et al., 2022;Li et al., 2021;Pamungkas, 2019). However, it is unclear if emotions from an AI would have the same effect. ...

Towards an Online Empathetic Chatbot with Emotion Causes
  • Citing Conference Paper
  • July 2021