Xiangnan Kong’s research while affiliated with Worcester Polytechnic Institute and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (147)


End-to-End Deep Learning for Structural Brain Imaging: A Unified Framework
  • Preprint

February 2025

·

3 Reads

Yao Su

·

·

Mingjie Zeng

·

[...]

·

Xiangnan Kong

Brain imaging analysis is fundamental in neuroscience, providing valuable insights into brain structure and function. Traditional workflows follow a sequential pipeline-brain extraction, registration, segmentation, parcellation, network generation, and classification-treating each step as an independent task. These methods rely heavily on task-specific training data and expert intervention to correct intermediate errors, making them particularly burdensome for high-dimensional neuroimaging data, where annotations and quality control are costly and time-consuming. We introduce UniBrain, a unified end-to-end framework that integrates all processing steps into a single optimization process, allowing tasks to interact and refine each other. Unlike traditional approaches that require extensive task-specific annotations, UniBrain operates with minimal supervision, leveraging only low-cost labels (i.e., classification and extraction) and a single labeled atlas. By jointly optimizing extraction, registration, segmentation, parcellation, network generation, and classification, UniBrain enhances both accuracy and computational efficiency while significantly reducing annotation effort. Experimental results demonstrate its superiority over existing methods across multiple tasks, offering a more scalable and reliable solution for neuroimaging analysis. Our code and data can be found at https://github.com/Anonymous7852/UniBrain



SkipSNN: Efficiently Classifying Spike Trains with Event-attention

October 2024

·

9 Reads

Spike train classification has recently become an important topic in the machine learning community, where each spike train is a binary event sequence with \emph{temporal-sparsity of signals of interest} and \emph{temporal-noise} properties. A promising model for it should follow the design principle of performing intensive computation only when signals of interest appear. So such tasks use mainly Spiking Neural Networks (SNNs) due to their consideration of temporal-sparsity of spike trains. However, the basic mechanism of SNNs ignore the temporal-noise issue, which makes them computationally expensive and thus high power consumption for analyzing spike trains on resource-constrained platforms. As an event-driven model, an SNN neuron makes a reaction given any input signals, making it difficult to quickly find signals of interest. In this paper, we introduce an event-attention mechanism that enables SNNs to dynamically highlight useful signals of the original spike trains. To this end, we propose SkipSNN, which extends existing SNN models by learning to mask out noise by skipping membrane potential updates and shortening the effective size of the computational graph. This process is analogous to how people choose to open and close their eyes to filter the information they see. We evaluate SkipSNN on various neuromorphic tasks and demonstrate that it achieves significantly better computational efficiency and classification accuracy than other state-of-the-art SNNs.


Amalgamating Multi-Task Models with Heterogeneous Architectures

March 2024

·

5 Reads

Proceedings of the AAAI Conference on Artificial Intelligence

Multi-task learning (MTL) is essential for real-world applications that handle multiple tasks simultaneously, such as selfdriving cars. MTL methods improve the performance of all tasks by utilizing information across tasks to learn a robust shared representation. However, acquiring sufficient labeled data tends to be extremely expensive, especially when having to support many tasks. Recently, Knowledge Amalgamation (KA) has emerged as an effective strategy for addressing the lack of labels by instead learning directly from pretrained models (teachers). KA learns one unified multi-task student that masters all tasks across all teachers. Existing KA for MTL works are limited to teachers with identical architectures, and thus propose layer-to-layer based approaches. Unfortunately, in practice, teachers may have heterogeneous architectures; their layers may not be aligned and their dimensionalities or scales may be incompatible. Amalgamating multi-task teachers with heterogeneous architectures remains an open problem. For this, we design Versatile Common Feature Consolidator (VENUS), the first solution to this problem. VENUS fuses knowledge from the shared representations of each teacher into one unified generalized representation for all tasks. Specifically, we design the Feature Consolidator network that leverages an array of teacher-specific trainable adaptors. These adaptors enable the student to learn from multiple teachers, even if they have incompatible learned representations. We demonstrate that VENUS outperforms five alternative methods on numerous benchmark datasets across a broad spectrum of experiments.




One-shot Joint Extraction, Registration and Segmentation of Neuroimaging Data

July 2023

·

19 Reads

·

1 Citation

Brain extraction, registration and segmentation are indispensable preprocessing steps in neuroimaging studies. The aim is to extract the brain from raw imaging scans (i.e., extraction step), align it with a target brain image (i.e., registration step) and label the anatomical brain regions (i.e., segmentation step). Conventional studies typically focus on developing separate methods for the extraction, registration and segmentation tasks in a supervised setting. The performance of these methods is largely contingent on the quantity of training samples and the extent of visual inspections carried out by experts for error correction. Nevertheless, collecting voxel-level labels and performing manual quality control on high-dimensional neuroimages (e.g., 3D MRI) are expensive and time-consuming in many medical studies. In this paper, we study the problem of one-shot joint extraction, registration and segmentation in neuroimaging data, which exploits only one labeled template image (a.k.a. atlas) and a few unlabeled raw images for training. We propose a unified end-to-end framework, called JERS, to jointly optimize the extraction, registration and segmentation tasks, allowing feedback among them. Specifically, we use a group of extraction, registration and segmentation modules to learn the extraction mask, transformation and segmentation mask, where modules are interconnected and mutually reinforced by self-supervision. Empirical results on real-world datasets demonstrate that our proposed method performs exceptionally in the extraction, registration and segmentation tasks. Our code and data can be found at https://github.com/Anonymous4545/JERS


Knowledge Amalgamation for Multi-Label Classification via Label Dependency Transfer

June 2023

·

8 Reads

·

4 Citations

Proceedings of the AAAI Conference on Artificial Intelligence

Multi-label classification (MLC), which assigns multiple labels to each instance, is crucial to domains from computer vision to text mining. Conventional methods for MLC require huge amounts of labeled data to capture complex dependencies between labels. However, such labeled datasets are expensive, or even impossible, to acquire. Worse yet, these pre-trained MLC models can only be used for the particular label set covered in the training data. Despite this severe limitation, few methods exist for expanding the set of labels predicted by pre-trained models. Instead, we acquire vast amounts of new labeled data and retrain a new model from scratch. Here, we propose combining the knowledge from multiple pre-trained models (teachers) to train a new student model that covers the union of the labels predicted by this set of teachers. This student supports a broader label set than any one of its teachers without using labeled data. We call this new problem knowledge amalgamation for multi-label classification. Our new method, Adaptive KNowledge Transfer (ANT), trains a student by learning from each teacher’s partial knowledge of label dependencies to infer the global dependencies between all labels across the teachers. We show that ANT succeeds in unifying label dependencies among teachers, outperforming five state-of-the-art methods on eight real-world datasets.


Finding Short Signals in Long Irregular Time Series with Continuous-Time Attention Policy Networks
  • Preprint
  • File available

February 2023

·

45 Reads

Irregularly-sampled time series (ITS) are native to high-impact domains like healthcare, where measurements are collected over time at uneven intervals. However, for many classification problems, only small portions of long time series are often relevant to the class label. In this case, existing ITS models often fail to classify long series since they rely on careful imputation, which easily over- or under-samples the relevant regions. Using this insight, we then propose CAT, a model that classifies multivariate ITS by explicitly seeking highly-relevant portions of an input series' timeline. CAT achieves this by integrating three components: (1) A Moment Network learns to seek relevant moments in an ITS's continuous timeline using reinforcement learning. (2) A Receptor Network models the temporal dynamics of both observations and their timing localized around predicted moments. (3) A recurrent Transition Model models the sequence of transitions between these moments, cultivating a representation with which the series is classified. Using synthetic and real data, we find that CAT outperforms ten state-of-the-art methods by finding short signals in long irregular time series.

Download

ABN: Anti-Blur Neural Networks for Multi-Stage Deformable Image Registration

December 2022

·

10 Reads

Deformable image registration, i.e., the task of aligning multiple images into one coordinate system by non-linear transformation, serves as an essential preprocessing step for neuroimaging data. Recent research on deformable image registration is mainly focused on improving the registration accuracy using multi-stage alignment methods, where the source image is repeatedly deformed in stages by a same neural network until it is well-aligned with the target image. Conventional methods for multi-stage registration can often blur the source image as the pixel/voxel values are repeatedly interpolated from the image generated by the previous stage. However, maintaining image quality such as sharpness during image registration is crucial to medical data analysis. In this paper, we study the problem of anti-blur deformable image registration and propose a novel solution, called Anti-Blur Network (ABN), for multi-stage image registration. Specifically, we use a pair of short-term registration and long-term memory networks to learn the nonlinear deformations at each stage, where the short-term registration network learns how to improve the registration accuracy incrementally and the long-term memory network combines all the previous deformations to allow an interpolation to perform on the raw image directly and preserve image sharpness. Extensive experiments on both natural and medical image datasets demonstrated that ABN can accurately register images while preserving their sharpness. Our code and data can be found at https://github.com/anonymous3214/ABN


Citations (67)


... In the literature, related tasks in brain imaging analysis have been extensively studied. Conventional methods primarily focus on designing methods for brain extraction (Kleesiek et al. 2016;Lucena et al. 2019), registration (Sokooti et al. 2017;Su et al. 2022a), segmentation (Akkus et al. 2017;Kamnitsas et al. 2017;Chen et al. 2018), parcellation (Thyreau and Taki 2020;Lim et al. 2022Lim et al. ), network generation (Škoch et al. 2022Yin et al. 2023) and classification Kawahara et al. 2017;Kan et al. 2022b) separately under supervised settings. However, in brain imaging studies, the collection of voxel-level annotations, transformations between images, and task-specific brain networks often prove to be expensive, as it demands extensive expertise, effort, and time to produce accurate labels, especially for high-dimensional neuroimaging data, e.g., 3D MRI. ...

Reference:

End-to-End Deep Learning for Structural Brain Imaging: A Unified Framework
Multi-State Brain Network Discovery
  • Citing Conference Paper
  • December 2023

... In [14], the issue of maximizing multi-item influence in continuous settings is examined, taking into account situations in which various influencers are offered disparate incentives on various items to entice them to participate in the viral marketing process. The goal of the cluster greedy algorithm, which is covered in [15], is to maximize the influence by dividing the social network into clusters and choosing seed influencers in an effective manner by combining the basic greedy algorithm with an investigation of the sub modularity property of the diffusion function. In [16], a quantum computing strategy for influence maximization is examined with the goal of obtaining near-optimal solutions through the use of effective quadratic unconstrained binary optimization formulations on quantum annealer and the transformation of the influence maximization problem into a max-cover instance problem. ...

Multi-Item Continuous Influence Maximization
  • Citing Conference Paper
  • December 2023

... A study was conducted on capturing the interdependence of labels in multiple-label classification, where an example can be assigned multiple labels simultaneously. This study also demonstrated that effectively managing the complexities associated with labels necessitates the use of advanced techniques, particularly in cases where certain labels are limited or require additional contextual information for accurate classification [35]. Nevertheless, employing techniques like synthetic data generation has been demonstrated to improve the performance of the model when dealing with imbalanced and diverse labels. ...

Knowledge Amalgamation for Multi-Label Classification via Label Dependency Transfer
  • Citing Article
  • June 2023

Proceedings of the AAAI Conference on Artificial Intelligence

... MetaST [26] employs a global memory queried by the target region. Moreover, STrans-GAN [31] generates future trac speed using GANs, and TPB method [15] proposes a trac pattern bank to store similar patterns from multiple source cities for the downstream ne-tuning task. However, these methods heavily depend on data-rich multiple source cities, and can be cost-prohibitive in practice. ...

STrans-GAN: Spatially-Transferable Generative Adversarial Networks for Urban Traffic Estimation
  • Citing Conference Paper
  • November 2022

... In the literature, related tasks in brain imaging analysis have been extensively studied. Conventional methods primarily focus on designing methods for brain extraction (Kleesiek et al. 2016;Lucena et al. 2019), registration (Sokooti et al. 2017;Su et al. 2022a), segmentation (Akkus et al. 2017;Kamnitsas et al. 2017;Chen et al. 2018), parcellation (Thyreau and Taki 2020;Lim et al. 2022Lim et al. ), network generation (Škoch et al. 2022Yin et al. 2023) and classification Kawahara et al. 2017;Kan et al. 2022b) separately under supervised settings. However, in brain imaging studies, the collection of voxel-level annotations, transformations between images, and task-specific brain networks often prove to be expensive, as it demands extensive expertise, effort, and time to produce accurate labels, especially for high-dimensional neuroimaging data, e.g., 3D MRI. ...

ABN: Anti-Blur Neural Networks for Multi-Stage Deformable Image Registration
  • Citing Conference Paper
  • November 2022

... Early classification of time series [1,2,3,4] is a pivotal algorithm, especially when sampling cost is high, e.g., medical early diagnosis [5], autonomous driving [6], and action recognition [7]. Under these applications, the early classifier seeks to optimize both speed and accuracy at the same time. ...

Stop&Hop: Early Classification of Irregular Time Series
  • Citing Article
  • October 2022

... If prediction is triggered, the hidden representation given by the RNN is sent to a Discriminator, whose role is to predict a class, given this representation. The model has been adapted to deal with irregularly sampled time series [57]. [58] extend the ECTS framework to channel filtering, using here also Reinforcement Learning. ...

Stop&Hop: Early Classification of Irregular Time Series
  • Citing Conference Paper
  • October 2022

... Two types of neural networks with recurrent LSTM layers [9] and LSTNet [10] were built to predict electricity imbalances. After which two linear layers with hyperbolic tangent and sigmoid activation functions, respectively. ...

Semi-Supervised Knowledge Amalgamation for Sequence Classification
  • Citing Article
  • May 2021

Proceedings of the AAAI Conference on Artificial Intelligence

... Instead of tedious, step-by-step processing for brain imaging data, recent studies support transforming these pipelines into deep neural networks for joint learning and end-to-end optimization (Ren et al. 2024;Agarwal et al. 2022). While several approaches have been proposed-such as joint extraction and registration (Su et al. 2022b), joint registration and parcellation (Zhao et al. 2021;Lord et al. 2007), and joint network generation and disease prediction (Campbell et al. 2022;Mahmood et al. 2021;Kan et al. 2022a)-there is currently no framework that unifies and simultaneously optimizes all these processing stages to directly create brain networks from raw imaging data. Mapping the connectome of human brain as a brain network (i.e., graph), has become one of the most pervasive paradigms in neuroscience (Sporns, Tononi, and Kotter 2005;Bargmann and Marder 2013). ...

ERNet: Unsupervised Collective Extraction and Registration in Neuroimaging Data
  • Citing Conference Paper
  • August 2022