Guangyi Zhang’s research while affiliated with Queen's University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (17)


Partial Label Learning for Emotion Recognition from EEG
  • Article

January 2025

IEEE Transactions on Affective Computing

Guangyi Zhang

·

Ali Etemad

Fully supervised learning has recently achieved promising performance in various electroencephalography (EEG) learning tasks by training on large datasets with ground truth labels. However, labeling EEG data for affective experiments is challenging, as it can be difficult for participants to accurately distinguish between similar emotions, resulting in ambiguous labeling (reporting multiple emotions for one EEG instance). This notion could cause model performance degradation, as the ground truth is hidden within multiple candidate labels. To address this issue, Partial Label Learning (PLL) has been proposed to identify the ground truth from candidate labels during the training phase, and has shown good performance in the computer vision domain. However, PLL methods have not yet been adopted for EEG representation learning or implemented for emotion recognition tasks. In this paper, we adapt and re-implement six state-of-the-art PLL approaches for emotion recognition from EEG on two large emotion datasets (SEED-IV and SEED-V). These datasets contain four and five categories of emotions, respectively. We evaluate the performance of all methods in classical, circumplex-based and real-world experiments. The results show that PLL methods can achieve strong results in affective computing from EEG and achieve comparable performance to fully supervised learning. We also investigate the effect of label disambiguation, a key step in many PLL methods. The results show that in most cases, label disambiguation would benefit the model when the candidate labels are generated based on their similarities to the ground truth rather than obeying a uniform distribution. This finding suggests the potential of using label disambiguation-based PLL methods for circumplex-based and real-world affective tasks.



Partial Label Learning for Emotion Recognition from EEG

February 2023

·

33 Reads

Fully supervised learning has recently achieved promising performance in various electroencephalography (EEG) learning tasks by training on large datasets with ground truth labels. However, labeling EEG data for affective experiments is challenging, as it can be difficult for participants to accurately distinguish between similar emotions, resulting in ambiguous labeling (reporting multiple emotions for one EEG instance). This notion could cause model performance degradation, as the ground truth is hidden within multiple candidate labels. To address this issue, Partial Label Learning (PLL) has been proposed to identify the ground truth from candidate labels during the training phase, and has shown good performance in the computer vision domain. However, PLL methods have not yet been adopted for EEG representation learning or implemented for emotion recognition tasks. In this paper, we adapt and re-implement six state-of-the-art PLL approaches for emotion recognition from EEG on a large emotion dataset (SEED-V, containing five emotion classes). We evaluate the performance of all methods in classical and real-world experiments. The results show that PLL methods can achieve strong results in affective computing from EEG and achieve comparable performance to fully supervised learning. We also investigate the effect of label disambiguation, a key step in many PLL methods. The results show that in most cases, label disambiguation would benefit the model when the candidate labels are generated based on their similarities to the ground truth rather than obeying a uniform distribution. This finding suggests the potential of using label disambiguation-based PLL methods for real-world affective tasks. We make the source code of this paper publicly available at: https://github.com/guangyizhangbci/PLL-Emotion-EEG.


Spatio-Temporal EEG Representation Learning on Riemannian Manifold and Euclidean Space

January 2023

·

15 Reads

·

24 Citations

IEEE Transactions on Emerging Topics in Computational Intelligence

We present a novel deep neural architecture for learning electroencephalogram (EEG). To learn the spatial information, our model first obtains the Riemannian mean and distance from spatial covariance matrices (SCMs) on a Riemannian manifold. We then project the spatial information onto a Euclidean space via tangent space learning. Following, two fully connected layers are used to learn the spatial information embeddings. Moreover, our proposed method learns the temporal information via differential entropy and logarithm power spectrum density features extracted from EEG signals in a Euclidean space using a deep long short-term memory network with a soft attention mechanism. To combine the spatial and temporal information, we use an effective fusion strategy, which learns attention weights applied to embedding-specific features for decision making. We evaluate our proposed framework on four public datasets across three popular EEG-related tasks, notably emotion recognition, vigilance estimation, and motor imagery classification, containing various types of tasks such as binary classification, multi-class classification, and regression. Our proposed architecture outperforms other methods on SEED-VIG, and approaches the state-of-the-art on the other three datasets (SEED, BCI-IV 2 A, and BCI-IV 2B), showing the robustness of our framework in EEG representation learning.


PARSE: Pairwise Alignment of Representations in Semi-Supervised EEG Learning for Emotion Recognition

October 2022

·

20 Reads

·

52 Citations

IEEE Transactions on Affective Computing

We propose PARSE, a novel semi-supervised architecture for learning reliable EEG representations for emotion recognition. To reduce the potential distribution mismatch between large amounts of unlabeled data and a limited number of labeled data, PARSE uses pairwise representation alignment. First, our model performs data augmentation followed by label guessing for large amounts of original and augmented unlabeled data. The model is then followed by sharpening the guessed labels and convex combinations of the unlabeled and labeled data. Finally, it performs representation alignment and emotion classification. To rigorously test our model, we compare PARSE to several state-of-the-art semi-supervised approaches, which we implement and adapt for EEG learning. We perform these experiments on four public EEG-based emotion recognition datasets, SEED, SEED-IV, SEED-V and AMIGOS (valence and arousal). The experiments show that our proposed framework achieves the overall best results with varying amounts of limited labeled samples in SEED, SEED-IV and AMIGOS (valence), while approaching the overall best result (reaching the second-best) in SEED-V and AMIGOS (arousal). The analysis shows that our pairwise representation alignment considerably improves the performance by performing the distribution alignment between unlabeled and labeled data, especially when only 1 sample per class is labeled. The source code of our paper is publicly available at https://github.com/guangyizhangbci/PARSE .



PARSE: Pairwise Alignment of Representations in Semi-Supervised EEG Learning for Emotion Recognition

February 2022

·

21 Reads

We propose PARSE, a novel semi-supervised architecture for learning strong EEG representations for emotion recognition. To reduce the potential distribution mismatch between the large amounts of unlabeled data and the limited amount of labeled data, PARSE uses pairwise representation alignment. First, our model performs data augmentation followed by label guessing for large amounts of original and augmented unlabeled data. This is then followed by sharpening of the guessed labels and convex combinations of the unlabeled and labeled data. Finally, representation alignment and emotion classification are performed. To rigorously test our model, we compare PARSE to several state-of-the-art semi-supervised approaches which we implement and adapt for EEG learning. We perform these experiments on four public EEG-based emotion recognition datasets, SEED, SEED-IV, SEED-V and AMIGOS (valence and arousal). The experiments show that our proposed framework achieves the overall best results with varying amounts of limited labeled samples in SEED, SEED-IV and AMIGOS (valence), while approaching the overall best result (reaching the second-best) in SEED-V and AMIGOS (arousal). The analysis shows that our pairwise representation alignment considerably improves the performance by reducing the distribution alignment between unlabeled and labeled data, especially when only 1 sample per class is labeled.



Holistic Semi-Supervised Approaches for EEG Representation Learning

September 2021

·

64 Reads

Recently, supervised methods, which often require substantial amounts of class labels, have achieved promising results for EEG representation learning. However, labeling EEG data is a challenging task. More recently, holistic semi-supervised learning approaches, which only require few output labels, have shown promising results in the field of computer vision. These methods, however, have not yet been adapted for EEG learning. In this paper, we adapt three state-of-the-art holistic semi-supervised approaches, namely MixMatch, FixMatch, and AdaMatch, as well as five classical semi-supervised methods for EEG learning. We perform rigorous experiments with all 8 methods on two public EEG-based emotion recognition datasets, namely SEED and SEED-IV. The experiments with different amounts of limited labeled samples show that the holistic approaches achieve strong results even when only 1 labeled sample is used per class. Further experiments show that in most cases, AdaMatch is the most effective method, followed by MixMatch and FixMatch.


Deep Recurrent Semi-Supervised EEG Representation Learning for Emotion Recognition

July 2021

·

39 Reads

EEG-based emotion recognition often requires sufficient labeled training samples to build an effective computational model. Labeling EEG data, on the other hand, is often expensive and time-consuming. To tackle this problem and reduce the need for output labels in the context of EEG-based emotion recognition, we propose a semi-supervised pipeline to jointly exploit both unlabeled and labeled data for learning EEG representations. Our semi-supervised framework consists of both unsupervised and supervised components. The unsupervised part maximizes the consistency between original and reconstructed input data using an autoencoder, while simultaneously the supervised part minimizes the cross-entropy between the input and output labels. We evaluate our framework using both a stacked autoencoder and an attention-based recurrent autoencoder. We test our framework on the large-scale SEED EEG dataset and compare our results with several other popular semi-supervised methods. Our semi-supervised framework with a deep attention-based recurrent autoencoder consistently outperforms the benchmark methods, even when small sub-sets (3\%, 5\% and 10\%) of the output labels are available during training, achieving a new state-of-the-art semi-supervised performance.


Citations (9)


... For example, Zhang et al. introduced the concept of manifolds, proposing a time-frequency domain feature learning model that integrated both Riemannian manifold and Euclidean space representations. Their work demonstrated the effectiveness of attention mechanisms in synthesizing feature information across different mathematical domains [39]. Additionally, to better capture the spatial relationships among EEG electrode channels, Jia et al. 's GraphSleepNet [40] and Zhang et al. 's hierarchical attention network based on graph structures [41] both utilized GCNs to model the spatial relationships of EEG electrodes. ...

Reference:

Integrating Biological and Machine Intelligence: Attention Mechanisms in Brain-Computer Interfaces
Spatio-Temporal EEG Representation Learning on Riemannian Manifold and Euclidean Space
  • Citing Article
  • January 2023

IEEE Transactions on Emerging Topics in Computational Intelligence

... For instance, FLDNet [14] enhances EEG emotion recognition by internally distilling features across multiple neural network layers, utilizing a triple-net structure and a frame gate layer to refine and integrate semantic information from various frames. Similarly, MLBNet [15] uses a Deep Mutual Learning (DML) strategy, where two networks specializing in temporal and spatial features mutually enhance their learning, thereby improving accuracy and adaptation to complex data. ...

Distilling EEG Representations via Capsules for Affective Computing
  • Citing Article
  • May 2023

Pattern Recognition Letters

... In the same order of ideas, [16] completed a five primary emotions classification by computing 18 EC using brain functional connectivity patterns per primary emotion, proposing an accurate multimodal approach. Alternatively, [17] developed a robust semi-supervised classification method by evaluating the clustering performance for S τ = {C τ |∀τ = {1, 3, 5, 7, 10, 25}} clusters and using primary and primitive datasets. Clustering showed a significant performance by increasing the labels in the primary emotions dataset; meanwhile, the primitive dataset showed a lesser sensitive performance by increasing the number of labels per cluster. ...

PARSE: Pairwise Alignment of Representations in Semi-Supervised EEG Learning for Emotion Recognition
  • Citing Article
  • October 2022

IEEE Transactions on Affective Computing

... The progressed advanced SSL methods were employed by Zhang et al [11], from machine vision for EEG-based emotion recognition performance, a technique rarely studied in SSL settings for EEG-based tasks. The authors [12], used naïve Gaussian noise as an augmentation [13], strategy for EEG data, using the mix-up augmentation technique. ...

Holistic Semi-Supervised Approaches for EEG Representation Learning
  • Citing Conference Paper
  • May 2022

... However, the controlled setup of the study makes it impractical to apply findings to real-world driving conditions. EEG and EOG recordings [13] were utilized to estimate the level of vigilance inside the vehicle. Long short-term memory (LSTM) combined with capsule feature extraction was employed to learn representative features. ...

Capsule Attention for Multimodal EEG-EOG Representation Learning With Application to Driver Vigilance Estimation
  • Citing Article
  • Full-text available
  • June 2021

IEEE transactions on neural systems and rehabilitation engineering: a publication of the IEEE Engineering in Medicine and Biology Society

... For evaluation, the Pearson Correlation Coefficient (PCC) and Root Mean Squared Error (RMSE) are employed as evaluation metrics for this task [54]. PCC provides the linear relationship between prediction and ground truth, taking on a value within the range of [-1, 1]. ...

RFNet: Riemannian Fusion Network for EEG-based Brain-Computer Interfaces
  • Citing Preprint
  • August 2020

... However, LSTM processes information sequentially and relies on the last hidden state, which limits its ability to focus on critical time points. To address this limitation, we incorporated an attention mechanism into the LSTM framework [28][29][30], allowing the model to effectively learn complex patterns by emphasizing important temporal features. ...

Classification of Hand Movements from EEG using a Deep Attention-based LSTM Network
  • Citing Article
  • December 2019

IEEE Sensors Journal

Guangyi Zhang

·

Vandad Davoodnia

·

·

[...]

·

Ali Etemad

... In this study, good accuracy was obtained in four-stage fatigue classification by using an end-to-end deep learning method with optimal deep layers to help the model learn detailed features and avoid maximum voluntary contraction (MVC)based measurements, which cannot always be performed in clinical settings. A study that employed MVC-based upper limb isometric contraction to detect low-level fatigue by using wavelet decomposition based on selected features was performed with an accuracy of 83% [19]. Another study that was performed based on EMG signals in cyclo-ergometric exercise by using an SVM for a set of nine features obtained an accuracy of 82% [8]. ...

Non-invasive detection of low-level muscle fatigue using surface EMG with wavelet decomposition
  • Citing Conference Paper
  • July 2018

Conference proceedings: ... Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Conference