Kyumin Lee’s research while affiliated with Worcester Polytechnic Institute and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (84)


Figure 1. Comparison of traditional Machine Learning approach v.s. our multi-modality alignment approach. Traditional approaches directly map labels into binary values. In contrast, our approach leverages the language model to encode the semantic relationship between context and activities within highdimensional vector representations and leads to better performance.
Figure 3. The SEAL framework consists of three main components: a Sensor Data Encoder, a CA-HAR Label Encoder, and a Modal Alignment. The Sensor Data Encoder transforms input sensor data into vector embedding representations, while the Label Encoder generates semantic label vector embedding representations from tokenized label sentences. Finally, the Modal Alignment component aligns the sensor data and CA-HAR label representations by maximizing their similarity, enabling SEAL to make accurate predictions. "Trm" are transformers modules within language models.
Figure 5. UMAP visualization of SEAL learned text embedding across all datasets, with clusters generated by KMeans. The clusters indicate SEAL's ability to capture the semantic relationship between activities and contexts.
Semantically Encoding Activity Labels for Context-Aware Human Activity Recognition
  • Preprint
  • File available

April 2025

·

2 Reads

·

Guanyi Mou

·

Emmanuel O. Agu

·

Kyumin Lee

Prior work has primarily formulated CA-HAR as a multi-label classification problem, where model inputs are time-series sensor data and target labels are binary encodings representing whether a given activity or context occurs. These CA-HAR methods either predicted each label independently or manually imposed relationships using graphs. However, both strategies often neglect an essential aspect: activity labels have rich semantic relationships. For instance, walking, jogging, and running activities share similar movement patterns but differ in pace and intensity, indicating that they are semantically related. Consequently, prior CA-HAR methods often struggled to accurately capture these inherent and nuanced relationships, particularly on datasets with noisy labels typically used for CA-HAR or situations where the ideal sensor type is unavailable (e.g., recognizing speech without audio sensors). To address this limitation, we propose SEAL, which leverage LMs to encode CA-HAR activity labels to capture semantic relationships. LMs generate vector embeddings that preserve rich semantic information from natural language. Our SEAL approach encodes input-time series sensor data from smart devices and their associated activity and context labels (text) as vector embeddings. During training, SEAL aligns the sensor data representations with their corresponding activity/context label embeddings in a shared embedding space. At inference time, SEAL performs a similarity search, returning the CA-HAR label with the embedding representation closest to the input data. Although LMs have been widely explored in other domains, surprisingly, their potential in CA-HAR has been underexplored, making our approach a novel contribution to the field. Our research opens up new possibilities for integrating more advanced LMs into CA-HAR tasks.

Download




From Critique to Clarity: A Pathway to Faithful and Personalized Code Explanations with Large Language Models

December 2024

·

13 Reads

Zexing Xu

·

Zhuang Luo

·

Yichuan Li

·

[...]

·

S. Rasoul Etesami

In the realm of software development, providing accurate and personalized code explanations is crucial for both technical professionals and business stakeholders. Technical professionals benefit from enhanced understanding and improved problem-solving skills, while business stakeholders gain insights into project alignments and transparency. Despite the potential, generating such explanations is often time-consuming and challenging. This paper presents an innovative approach that leverages the advanced capabilities of large language models (LLMs) to generate faithful and personalized code explanations. Our methodology integrates prompt enhancement, self-correction mechanisms, personalized content customization, and interaction with external tools, facilitated by collaboration among multiple LLM agents. We evaluate our approach using both automatic and human assessments, demonstrating that our method not only produces accurate explanations but also tailors them to individual user preferences. Our findings suggest that this approach significantly improves the quality and relevance of code explanations, offering a valuable tool for developers and stakeholders alike.


Contrastive Learning with Auxiliary User Detection for Identifying Activities

October 2024

·

7 Reads

Human Activity Recognition (HAR) is essential in ubiquitous computing, with far-reaching real-world applications. While recent SOTA HAR research has demonstrated impressive performance, some key aspects remain under-explored. Firstly, HAR can be both highly contextualized and personalized. However, prior work has predominantly focused on being Context-Aware (CA) while largely ignoring the necessity of being User-Aware (UA). We argue that addressing the impact of innate user action-performing differences is equally crucial as considering external contextual environment settings in HAR tasks. Secondly, being user-aware makes the model acknowledge user discrepancies but does not necessarily guarantee mitigation of these discrepancies, i.e., unified predictions under the same activities. There is a need for a methodology that explicitly enforces closer (different user, same activity) representations. To bridge this gap, we introduce CLAUDIA, a novel framework designed to address these issues. Specifically, we expand the contextual scope of the CA-HAR task by integrating User Identification (UI) within the CA-HAR framework, jointly predicting both CA-HAR and UI in a new task called User and Context-Aware HAR (UCA-HAR). This approach enriches personalized and contextual understanding by jointly learning user-invariant and user-specific patterns. Inspired by SOTA designs in the visual domain, we introduce a supervised contrastive loss objective on instance-instance pairs to enhance model efficacy and improve learned feature quality. Evaluation across three real-world CA-HAR datasets reveals substantial performance enhancements, with average improvements ranging from 5.8% to 14.1% in Matthew's Correlation Coefficient and 3.0% to 7.2% in Macro F1 score.


Let's Ask GNN: Empowering Large Language Model for Graph In-Context Learning

October 2024

·

7 Reads

Textual Attributed Graphs (TAGs) are crucial for modeling complex real-world systems, yet leveraging large language models (LLMs) for TAGs presents unique challenges due to the gap between sequential text processing and graph-structured data. We introduce AskGNN, a novel approach that bridges this gap by leveraging In-Context Learning (ICL) to integrate graph data and task-specific information into LLMs. AskGNN employs a Graph Neural Network (GNN)-powered structure-enhanced retriever to select labeled nodes across graphs, incorporating complex graph structures and their supervision signals. Our learning-to-retrieve algorithm optimizes the retriever to select example nodes that maximize LLM performance on graph. Experiments across three tasks and seven LLMs demonstrate AskGNN's superior effectiveness in graph task performance, opening new avenues for applying LLMs to graph-structured data without extensive fine-tuning.


Deep Heterogeneous Contrastive Hyper-Graph Learning for In-the-Wild Context-Aware Human Activity Recognition

September 2024

·

1 Read

Human Activity Recognition (HAR) is a challenging, multi-label classification problem as activities may co-occur and sensor signals corresponding to the same activity may vary in different contexts (e.g., different device placements). This paper proposes a Deep Heterogeneous Contrastive Hyper-Graph Learning (DHC-HGL) framework that captures heterogenous Context-Aware HAR (CA-HAR) hypergraph properties in a message-passing and neighborhood-aggregation fashion. Prior work only explored homogeneous or shallow-node-heterogeneous graphs. DHC-HGL handles heterogeneous CA-HAR data by innovatively 1) Constructing three different types of sub-hypergraphs that are each passed through different custom HyperGraph Convolution (HGC) layers designed to handle edge-heterogeneity and 2) Adopting a contrastive loss function to ensure node-heterogeneity. In rigorous evaluation on two CA-HAR datasets, DHC-HGL significantly outperformed state-of-the-art baselines by 5.8% to 16.7% on Matthews Correlation Coefficient (MCC) and 3.0% to 8.4% on Macro F1 scores. UMAP visualizations of learned CA-HAR node embeddings are also presented to enhance model explainability.


Fig. 2. An Overview of our HHGNN-CHAR Framework.
Heterogeneous Hyper-Graph Neural Networks for Context-aware Human Activity Recognition

September 2024

·

4 Reads

Context-aware Human Activity Recognition (CHAR) is challenging due to the need to recognize the user's current activity from signals that vary significantly with contextual factors such as phone placements and the varied styles with which different users perform the same activity. In this paper, we argue that context-aware activity visit patterns in realistic in-the-wild data can equivocally be considered as a general graph representation learning task. We posit that exploiting underlying graphical patterns in CHAR data can improve CHAR task performance and representation learning. Building on the intuition that certain activities are frequently performed with the phone placed in certain positions, we focus on the context-aware human activity problem of recognizing the <Activity, Phone Placement> tuple. We demonstrate that CHAR data has an underlying graph structure that can be viewed as a heterogenous hypergraph that has multiple types of nodes and hyperedges (an edge connecting more than two nodes). Subsequently, learning <Activity, Phone Placement> representations becomes a graph node representation learning problem. After task transformation, we further propose a novel Heterogeneous HyperGraph Neural Network architecture for Context-aware Human Activity Recognition (HHGNN-CHAR), with three types of heterogeneous nodes (user, phone placement, and activity). Connections between all types of nodes are represented by hyperedges. Rigorous evaluation demonstrated that on an unscripted, in-the-wild CHAR dataset, our proposed framework significantly outperforms state-of-the-art (SOTA) baselines including CHAR models that do not exploit graphs, and GNN variants that do not incorporate heterogeneous nodes or hyperedges with overall improvements 14.04% on Matthews Correlation Coefficient (MCC) and 7.01% on Macro F1 scores.


SWE2: SubWord Enriched and Significant Word Emphasized Framework for Hate Speech Detection

September 2024

·

9 Reads

Hate speech detection on online social networks has become one of the emerging hot topics in recent years. With the broad spread and fast propagation speed across online social networks, hate speech makes significant impacts on society by increasing prejudice and hurting people. Therefore, there are aroused attention and concern from both industry and academia. In this paper, we address the hate speech problem and propose a novel hate speech detection framework called SWE2, which only relies on the content of messages and automatically identifies hate speech. In particular, our framework exploits both word-level semantic information and sub-word knowledge. It is intuitively persuasive and also practically performs well under a situation with/without character-level adversarial attack. Experimental results show that our proposed model achieves 0.975 accuracy and 0.953 macro F1, outperforming 7 state-of-the-art baselines under no adversarial attack. Our model robustly and significantly performed well under extreme adversarial attack (manipulation of 50% messages), achieving 0.967 accuracy and 0.934 macro F1.


Citations (49)


... Yoo et al. (2021) applied LLMs to predict soft labels and then used knowledge distillation alongside text perturbation methods such as lexical substitution and word order shuffling to create augmented datasets. Addressing the challenge of adapting general instructions to a variety of downstream tasks, Li et al. (2024b) employed LLMs to generate a wide range of augmented instructions and implemented a scoring system to select task-relevant instructions, which improved model generalization. Additionally, Chen et al. (2024) focused on few-shot question answering, using graph algorithms and unsupervised question generation to extract the most informative data, thereby reducing the need for extensive fine-tuning data and enhancing training efficiency. ...

Reference:

Span-level Emotion-Cause-Category Triplet Extraction with Instruction Tuning LLMs and Data Augmentation
Empowering Large Language Models for Textual Data Augmentation
  • Citing Conference Paper
  • January 2024

... Verification labels are increasingly being used on social media platforms, including by journalists and news brands, to authenticate accounts (Vaidya et al. 2019). While these platforms refer to these blue check accounts as being more credible, scholars have argued that the label might be redundant, as they can be purchased (Edgerly and Vraga 2019;Varzgani et al. 2023;Wang et al. 2021). Nonetheless, in public debate, these blue check accounts are claimed to be more authentic and, hence, reliable (e.g. Warren 2023). ...

Toward Designing Effective Warning Labels for Health Misinformation on Social Media
  • Citing Conference Paper
  • January 2023

... HAR-GCNN [22] leveraged chronological correlations between sequential activities to predict missing labels. HHGNN [6] and DHC-HGL [7] proposed modeling the activity and context labels as nodes in a heterogeneous hypergraph. However, the manual definition of these graphs may restrict the expressiveness and adaptability of the learned representations. ...

Deep Heterogeneous Contrastive Hyper-Graph Learning for In-the-Wild Context-Aware Human Activity Recognition
  • Citing Article
  • January 2024

Proceedings of the ACM on Interactive Mobile Wearable and Ubiquitous Technologies

... Similarly, methods such as MoleculeSTM [104], CLAMP [105], ConGraT [106], G2P2 [107], GRENADE [108] also employ independent GNN encoders and LLMs encoders to process molecular and textual data separately and then map the embedded representations to a shared joint representation space for contrastive learning. However, these models differ in implementation details or in the specific application scenarios they are applicable to. ...

GRENADE: Graph-Centric Language Model for Self-Supervised Representation Learning on Text-Attributed Graphs
  • Citing Conference Paper
  • January 2023

... More recent work has explored modeling label co-occurrences using manually defined, learned graphical representations to capture some relationships between labels [6]- [8]. However, these approaches overlook a key consideration: activity labels have intrinsic, nuanced semantic relationships that are not captured when labels are assigned independent binary values, or missed by manually defined graphical relationships. ...

Heterogeneous Hyper-Graph Neural Networks for Context-Aware Human Activity Recognition
  • Citing Conference Paper
  • March 2023

... However, since all these supervised learning solutions need large enough amounts of updated labelled data to avoid overfitting/underfitting phenomena, they cannot deal appropriately with issue I1 (and, partly, with issue I2). Few proposals in the field addressed label-scarce settings by using a Weakly-Supervised approach leveraging Semi-Supervised Learning (SSL) [10][11][12][13][14][15] or Active Learning (AL) [16][17][18][19][20] methods. However, most of the proposed solutions were devised to take context (propagation and/or author/topic-related) information [3] as input. ...

Energy-based Domain Adaption with Active Learning for Emerging Misinformation Detection
  • Citing Conference Paper
  • December 2022

... Some research suggests that social media campaigns can encourage small actions, such as reducing the use of single-use plastics or saving energy. However, sustainable and long-term behaviour change may require a more comprehensive approach, including education, policies and incentives (Lee et al., 2021) . ...

Crowdturfers, Campaigns, and Social Media: Tracking and Revealing Crowdsourced Manipulation of Social Media
  • Citing Article
  • August 2021

Proceedings of the International AAAI Conference on Web and Social Media

... Research has demonstrated that digitalization has resulted in increased self-employment and entrepreneurship [18], requiring workers to develop new skills to work online [63,81]. In a study of musicians, Baym [9] has found that digitalization via social media meant that musicians felt more pressure to cultivate online relationships with fans who wanted direct contact while balancing it against their own needs for privacy. ...

Crowds, Gigs, and Super Sellers: A Measurement Study of a Supply-Driven Crowdsourcing Marketplace
  • Citing Article
  • August 2021

Proceedings of the International AAAI Conference on Web and Social Media

... Other studies use large-scale data, including LBSN data, and data mining techniques to understand which factors may be associated with people's movement patterns. For example, Cheng et al. [2021] used geolocated data from Twitter to understand user movements. The authors associated this spatial information with the economic characteristics of users, the geographic aspects of the areas frequented, as well as their positioning within the social network and the language used in their check-ins. ...

Exploring Millions of Footprints in Location Sharing Services
  • Citing Article
  • August 2021

Proceedings of the International AAAI Conference on Web and Social Media

... Various Machine-Learning methods have been trained to detect social bots with a combination of features extracted from the social network structure, content/profile attributes, and temporal patterns [19]. Alternatively, all of these feature types are combined into a single model [15,22,33,47,53,58]. Bot-detection algorithms often start by modeling the characteristics of accounts as the first step to distinguish bot from human-like behaviors [8]. ...

Seven Months with the Devils: A Long-Term Study of Content Polluters on Twitter
  • Citing Article
  • August 2021

Proceedings of the International AAAI Conference on Web and Social Media