Shenyang Huang’s research while affiliated with McGill University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (22)


Recommendation results (MAP, higher is better, in %) on RELBENCH.
Results on Amazon-Book.
Runtime [s] of 1,000 optimization steps.
ContextGNN: Beyond Two-Tower Recommendation Systems
  • Preprint
  • File available

November 2024

·

58 Reads

Yiwen Yuan

·

Zecheng Zhang

·

Xinwei He

·

[...]

·

Recommendation systems predominantly utilize two-tower architectures, which evaluate user-item rankings through the inner product of their respective embeddings. However, one key limitation of two-tower models is that they learn a pair-agnostic representation of users and items. In contrast, pair-wise representations either scale poorly due to their quadratic complexity or are too restrictive on the candidate pairs to rank. To address these issues, we introduce Context-based Graph Neural Networks (ContextGNNs), a novel deep learning architecture for link prediction in recommendation systems. The method employs a pair-wise representation technique for familiar items situated within a user's local subgraph, while leveraging two-tower representations to facilitate the recommendation of exploratory items. A final network then predicts how to fuse both pair-wise and two-tower recommendations into a single ranking of items. We demonstrate that ContextGNN is able to adapt to different data characteristics and outperforms existing methods, both traditional and GNN-based, on a diverse set of practical recommendation tasks, improving performance by 20% on average.

Download

Figure 4: Different setting for evaluation of future link prediction include between deployed, streaming and live-update setting. UTG framework is designed for the streaming setting.
Dataset statistics. Dataset # Nodes # Edges # Unique Edges Surprise Time Granularity # Snapshots
Test MRR comparison for snapshot and event-based methods on DTDG datasets, results reported from 5 runs. Top three models are coloured by First, Second, Third.
Test inference time comparison for snapshot and event based methods on DTDG datasets, we report the average result from 5 runs. Top three models are coloured by First, Second, Third.
Test inference time comparison for snapshot and event based methods on CTDG datasets, results reported from 5 runs. Top three models are coloured by First, Second, Third.
UTG: Towards a Unified View of Snapshot and Event Based Models for Temporal Graphs

July 2024

·

16 Reads

Temporal graphs have gained increasing importance due to their ability to model dynamically evolving relationships. These graphs can be represented through either a stream of edge events or a sequence of graph snapshots. Until now, the development of machine learning methods for both types has occurred largely in isolation, resulting in limited experimental comparison and theoretical crosspollination between the two. In this paper, we introduce Unified Temporal Graph (UTG), a framework that unifies snapshot-based and event-based machine learning models under a single umbrella, enabling models developed for one representation to be applied effectively to datasets of the other. We also propose a novel UTG training procedure to boost the performance of snapshot-based models in the streaming setting. We comprehensively evaluate both snapshot and event-based models across both types of temporal graphs on the temporal link prediction task. Our main findings are threefold: first, when combined with UTG training, snapshotbased models can perform competitively with event-based models such as TGN and GraphMixer even on event datasets. Second, snapshot-based models are at least an order of magnitude faster than most event-based models during inference. Third, while event-based methods such as NAT and DyGFormer outperforms snapshotbased methods on both types of temporal graphs, this is because they leverage joint neighborhood structural features thus emphasizing the potential to incorporate these features into snapshot-based models as well. These findings highlight the importance of comparing model architectures independent of the data format and suggest the potential of combining the efficiency of snapshot-based models with the performance of event-based models in the future.


Towards Neural Scaling Laws for Foundation Models on Temporal Graphs

June 2024

·

58 Reads

The field of temporal graph learning aims to learn from evolving network data to forecast future interactions. Given a collection of observed temporal graphs, is it possible to predict the evolution of an unseen network from the same domain? To answer this question, we first present the Temporal Graph Scaling (TGS) dataset, a large collection of temporal graphs consisting of eighty-four ERC20 token transaction networks collected from 2017 to 2023. Next, we evaluate the transferability of Temporal Graph Neural Networks (TGNNs) for the temporal graph property prediction task by pre-training on a collection of up to sixty-four token transaction networks and then evaluating the downstream performance on twenty unseen token networks. We find that the neural scaling law observed in NLP and Computer Vision also applies in temporal graph learning, where pre-training on greater number of networks leads to improved downstream performance. To the best of our knowledge, this is the first empirical demonstration of the transferability of temporal graphs learning. On downstream token networks, the largest pre-trained model outperforms single model TGNNs on thirteen unseen test networks. Therefore, we believe that this is a promising first step towards building foundation models for temporal graphs.


TGB 2.0: A Benchmark for Learning on Temporal Knowledge Graphs and Heterogeneous Graphs

June 2024

·

14 Reads

Multi-relational temporal graphs are powerful tools for modeling real-world data, capturing the evolving and interconnected nature of entities over time. Recently, many novel models are proposed for ML on such graphs intensifying the need for robust evaluation and standardized benchmark datasets. However, the availability of such resources remains scarce and evaluation faces added complexity due to reproducibility issues in experimental protocols. To address these challenges, we introduce Temporal Graph Benchmark 2.0 (TGB 2.0), a novel benchmarking framework tailored for evaluating methods for predicting future links on Temporal Knowledge Graphs and Temporal Heterogeneous Graphs with a focus on large-scale datasets, extending the Temporal Graph Benchmark. TGB 2.0 facilitates comprehensive evaluations by presenting eight novel datasets spanning five domains with up to 53 million edges. TGB 2.0 datasets are significantly larger than existing datasets in terms of number of nodes, edges, or timestamps. In addition, TGB 2.0 provides a reproducible and realistic evaluation pipeline for multi-relational temporal graphs. Through extensive experimentation, we observe that 1) leveraging edge-type information is crucial to obtain high performance, 2) simple heuristic baselines are often competitive with more complex methods, 3) most methods fail to run on our largest datasets, highlighting the need for research on more scalable methods.


Model Hyperparameters.
Temporal Graph Rewiring with Expander Graphs

June 2024

·

31 Reads

Evolving relations in real-world networks are often modelled by temporal graphs. Graph rewiring techniques have been utilised on Graph Neural Networks (GNNs) to improve expressiveness and increase model performance. In this work, we propose Temporal Graph Rewiring (TGR), the first approach for graph rewiring on temporal graphs. TGR enables communication between temporally distant nodes in a continuous time dynamic graph by utilising expander graph propagation to construct a message passing highway for message passing between distant nodes. Expander graphs are suitable candidates for rewiring as they help overcome the oversquashing problem often observed in GNNs. On the public tgbl-wiki benchmark, we show that TGR improves the performance of a widely used TGN model by a significant margin. Our code repository is accessible at https://anonymous.4open.science/r/TGR-254C.


Static graph approximations of dynamic contact networks for epidemic forecasting

May 2024

·

35 Reads

Epidemic modeling is essential in understanding the spread of infectious diseases like COVID-19 and devising effective intervention strategies to control them. Recently, network-based disease models have integrated traditional compartment-based modeling with real-world contact graphs and shown promising results. However, in an ongoing epidemic, future contact network patterns are not observed yet. To address this, we use aggregated static networks to approximate future contacts for disease modeling. The standard method in the literature concatenates all edges from a dynamic graph into one collapsed graph, called the full static graph. However, the full static graph often leads to severe overestimation of key epidemic characteristics. Therefore, we propose two novel static network approximation methods, DegMST and EdgeMST, designed to preserve the sparsity of real world contact network while remaining connected. DegMST and EdgeMST use the frequency of temporal edges and the node degrees respectively to preserve sparsity. Our analysis show that our models more closely resemble the network characteristics of the dynamic graph compared to the full static ones. Moreover, our analysis on seven real-world contact networks suggests EdgeMST yield more accurate estimations of disease dynamics for epidemic forecasting when compared to the standard full static method.



Laplacian Change Point Detection for Single and Multi-view Dynamic Graphs

November 2023

·

15 Reads

·

6 Citations

ACM Transactions on Knowledge Discovery from Data

Dynamic graphs are rich data structures that are used to model complex relationships between entities over time. In particular, anomaly detection in temporal graphs is crucial for many real world applications such as intrusion identification in network systems, detection of ecosystem disturbances and detection of epidemic outbreaks. In this paper, we focus on change point detection in dynamic graphs and address three main challenges associated with this problem: i) how to compare graph snapshots across time, ii) how to capture temporal dependencies, and iii) how to combine different views of a temporal graph. To solve the above challenges, we first propose Laplacian Anomaly Detection (LAD) which uses the spectrum of graph Laplacian as the low dimensional embedding of the graph structure at each snapshot. LAD explicitly models short term and long term dependencies by applying two sliding windows. Next, we propose MultiLAD, a simple and effective generalization of LAD to multi-view graphs. MultiLAD provides the first change point detection method for multi-view dynamic graphs. It aggregates the singular values of the normalized graph Laplacian from different views through the scalar power mean operation. Through extensive synthetic experiments, we show that i) LAD and MultiLAD are accurate and outperforms state-of-the-art baselines and their multi-view extensions by a large margin, ii) MultiLAD’s advantage over contenders significantly increases when additional views are available, and iii) MultiLAD is highly robust to noise from individual views. In five real world dynamic graphs, we demonstrate that LAD and MultiLAD identify significant events as top anomalies such as the implementation of government COVID-19 interventions which impacted the population mobility in multi-view traffic networks.


Towards Temporal Edge Regression: A Case Study on Agriculture Trade Between Nations

August 2023

·

42 Reads

Recently, Graph Neural Networks (GNNs) have shown promising performance in tasks on dynamic graphs such as node classification, link prediction and graph regression. However, few work has studied the temporal edge regression task which has important real-world applications. In this paper, we explore the application of GNNs to edge regression tasks in both static and dynamic settings, focusing on predicting food and agriculture trade values between nations. We introduce three simple yet strong baselines and comprehensively evaluate one static and three dynamic GNN models using the UN Trade dataset. Our experimental results reveal that the baselines exhibit remarkably strong performance across various settings, highlighting the inadequacy of existing GNNs. We also find that TGN outperforms other GNN models, suggesting TGN is a more appropriate choice for edge regression tasks. Moreover, we note that the proportion of negative edges in the training samples significantly affects the test performance. The companion source code can be found at: https://github.com/scylj1/GNN_Edge_Regression.


Temporal Graph Benchmark for Machine Learning on Temporal Graphs

July 2023

·

26 Reads

·

1 Citation

We present the Temporal Graph Benchmark (TGB), a collection of challenging and diverse benchmark datasets for realistic, reproducible, and robust evaluation of machine learning models on temporal graphs. TGB datasets are of large scale, spanning years in duration, incorporate both node and edge-level prediction tasks and cover a diverse set of domains including social, trade, transaction, and transportation networks. For both tasks, we design evaluation protocols based on realistic use-cases. We extensively benchmark each dataset and find that the performance of common models can vary drastically across datasets. In addition, on dynamic node property prediction tasks, we show that simple methods often achieve superior performance compared to existing temporal graph models. We believe that these findings open up opportunities for future research on temporal graphs. Finally, TGB provides an automated machine learning pipeline for reproducible and accessible temporal graph research, including data loading, experiment setup and performance evaluation. TGB will be maintained and updated on a regular basis and welcomes community feedback. TGB datasets, data loaders, example codes, evaluation setup, and leaderboards are publicly available at https://tgb.complexdatalab.com/ .


Citations (8)


... In this context, temporal network theory, network science metrics, and temporal graph mining tools can be leveraged to enhance the interpretability of TGL models. For instance, two works [40,11] analyze the temporal edge re-occurrence in the TGB datasets, highlighting that deep learning models may vary their performance due to the level of re-occurrence/novelty of the edges in the datasets, depending on their memorization or inductive reasoning capabilities. Dileo et al. [8] tested wellknown network science heuristics for link prediction on TGB. ...

Reference:

Network Science Meets AI: A Converging Frontier
Temporal Graph Analysis with TGX
  • Citing Conference Paper
  • March 2024

... To detect whether or not a power system has anomalies, we compare the global connectivity (i.e., the overall similarity of bus voltages) of the spatiotemporal graphs at different times. The global connectivity of a spatial graph can be quantified by the Fiedler value (i.e., the second smallest eigenvalue λ 2 ) of the generalized graph Laplacian matrix [16][17][18]. A larger λ 2 indicates that the entire graph has strong connectivity as all nodes are well-connected. ...

Laplacian Change Point Detection for Single and Multi-view Dynamic Graphs
  • Citing Article
  • November 2023

ACM Transactions on Knowledge Discovery from Data

... This approach enables scalable hallucination detection while maintaining performance. By utilizing tools such as the Density of States (DOS) and the kernel polynomial method (KPM) for approximating EigenScore (Huang et al., 2023b;Lin et al., 2014), we aim to enhance the efficiency of our analysis in the context of confabulations, which we will demonstrate empirically with EES and SeND. ...

Fast and Attributed Change Detection on Dynamic Graphs with Density of States
  • Citing Chapter
  • May 2023

Lecture Notes in Computer Science

... However, these three approaches are not adapted to CIL. NAS for CIL methods [21,20] are generally time-consuming and computationally expensive, as they explore a large space of possible network architectures and require training of candidate models to evaluate an architecture. They are not suitable for the practical case of a model adapting quickly to a dynamic environment, which is encountered in CIL. ...

Understanding Capacity Saturation in Incremental Learning
  • Citing Article
  • June 2021

... More details about the epidemic process in complex networks can be found at [14,15]. The Laplacian diffusion has been used recently to understand spatial disease transmission dynamics by researchers [16,17,18,19,20,21]. It is also used to characterize population mobility through a complex network. ...

Incorporating dynamic flight network in SEIR to model mobility between populations

Applied Network Science

... This encoder employs the same scaling strategy as the linear time encoder to the time differences but encodes them with sinusoidal functions, allowing us to isolate the impact of scaling from the underlying encoding technique. (Fowler, 2006;Huang et al., 2020). We list the details of the datasets in Appendix C. Following the setup of the unified dynamic graph learning library, DyGLib (Yu et al., 2023), we split the time span of an entire dataset into 70%/15%/15% for train/validation/test. ...

Laplacian Change Point Detection for Dynamic Graphs
  • Citing Conference Paper
  • August 2020

... While effective at mitigating catastrophic forgetting, they cannot fully preserve prior knowledge. Architecture-based methods introduce new parameters for each dataset and learn them independently, potentially avoiding knowledge loss (Xu and Zhu, 2018;Huang et al., 2019;Razdaibiedina et al., 2023). However, managing these additional parameters remains challenging. ...

Neural Architecture Search for Class-incremental Learning
  • Citing Preprint
  • September 2019