Jiaxuan You’s research while affiliated with Stanford University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (28)


Graph-based clinical recommender: Predicting specialists procedure orders using graph representation learning
  • Article

June 2023

·

29 Reads

·

3 Citations

Journal of Biomedical Informatics

Sajjad Fouladvand

·

Federico Reyes Gomez

·

Hamed Nilforoshan

·

[...]

·

Jonathan Chen

Objective: To determine whether graph neural network based models of electronic health records can predict specialty consultation care needs for endocrinology and hematology more accurately than the standard of care checklists and other conventional medical recommendation algorithms in the literature. Methods: Demand for medical expertise far outstrips supply, with tens of millions in the US alone with deficient access to specialty care. Rather than potentially months long delays to initiate diagnostic workup and medical treatment with a specialist, referring primary care supported by an automated recommender algorithm could anticipate and directly initiate patient evaluation that would otherwise be needed at subsequent a specialist appointment. We propose a novel graph representation learning approach with a heterogeneous graph neural network to model structured electronic health records and formulate recommendation/prediction of subsequent specialist orders as a link prediction problem. Results: Models are trained and assessed in two specialty care sites: endocrinology and hematology. Our experimental results show that our model achieves an 8% improvement in ROC-AUC for endocrinology (ROC-AUC = 0.88) and 5% improvement for hematology (ROC-AUC = 0.84) personalized procedure recommendations over prior medical recommender systems. These recommender algorithm approaches provide medical procedure recommendations for endocrinology referrals more effectively than manual clinical checklists (recommender: precision = 0.60, recall = 0.27, F1-score = 0.37) vs. (checklist: precision = 0.16, recall = 0.28, F1-score = 0.20), and similarly for hematology referrals (recommender: precision = 0.44, recall = 0.38, F1-score = 0.41) vs. (checklist: precision = 0.27, recall = 0.71, F1-score = 0.39). Conclusion: Embedding graph neural network models into clinical care can improve digital specialty consultation systems and expand the access to medical experience of prior similar cases.


Design choices in our search space
AutoTransfer: AutoML with Knowledge Transfer -- An Application to Graph Neural Networks
  • Preprint
  • File available

March 2023

·

100 Reads

AutoML has demonstrated remarkable success in finding an effective neural architecture for a given machine learning task defined by a specific dataset and an evaluation metric. However, most present AutoML techniques consider each task independently from scratch, which requires exploring many architectures, leading to high computational cost. Here we propose AutoTransfer, an AutoML solution that improves search efficiency by transferring the prior architectural design knowledge to the novel task of interest. Our key innovation includes a task-model bank that captures the model performance over a diverse set of GNN architectures and tasks, and a computationally efficient task embedding that can accurately measure the similarity among different tasks. Based on the task-model bank and the task embeddings, we estimate the design priors of desirable models of the novel task, by aggregating a similarity-weighted sum of the top-K design distributions on tasks that are similar to the task of interest. The computed design priors can be used with any AutoML search algorithm. We evaluate AutoTransfer on six datasets in the graph machine learning domain. Experiments demonstrate that (i) our proposed task embedding can be computed efficiently, and that tasks with similar embeddings have similar best-performing architectures; (ii) AutoTransfer significantly improves search efficiency with the transferred design priors, reducing the number of explored architectures by an order of magnitude. Finally, we release GNN-Bank-101, a large-scale dataset of detailed GNN training information of 120,000 task-model combinations to facilitate and inspire future research.

Download

Figure 2: Comparing precision-Recall curve of our proposed model with the baselines.
Performance of endocrinologist pro- cedure order prediction models.
Graph-Based Clinical Recommender: Predicting Specialists Procedure Orders using Graph Representation Learning

November 2022

·

70 Reads

Objective To determine whether graph neural network based models of electronic health records can predict specialty consultation care needs for endocrinology and hematology more accurately than the standard of care checklists and other conventional medical recommendation algorithms in the literature. Methods Demand for medical expertise far outstrips supply, with tens of millions in the US alone with deficient access to specialty care. Rather than potentially months long delays to initiate diagnostic workup and medical treatment with a specialist, referring primary care supported by an automated recommender algorithm could anticipate and directly initiate patient evaluation that would otherwise be needed at subsequent a specialist appointment. We propose a novel graph representation learning approach with a heterogeneous graph neural network to model structured electronic health records and formulate recommendation/prediction of subsequent specialist orders as a link prediction problem. Results Models are trained and assessed in two specialty care sites: endocrinology and hematology. Our experimental results show that our model achieves an 8% improvement in ROC-AUC for endocrinology (ROC-AUC=0.88) and 5% improvement for hematology (ROC-AUC=0.84) personalized procedure recommendations over prior medical recommender systems. These recommender algorithm approaches provide medical procedure recommendations for endocrinology referrals more effectively than manual clinical checklists (recommender: precision=0.60, recall=0.27, F1-score=0.37) vs. (checklist: precision=0.16, recall=0.28, F1-score=0.20), and similarly for hematology referrals (recommender: precision=0.44, recall=0.38, F1-score=0.41) vs. (checklist: precision=0.27, recall=0.71, F1-score=0.39). Conclusion Embedding graph neural network models into clinical care can improve digital specialty consultation systems and expand the access to medical experience of prior similar cases.


Efficient Automatic Machine Learning via Design Graphs

October 2022

·

50 Reads

Despite the success of automated machine learning (AutoML), which aims to find the best design, including the architecture of deep networks and hyper-parameters, conventional AutoML methods are computationally expensive and hardly provide insights into the relations of different model design choices. To tackle the challenges, we propose FALCON, an efficient sample-based method to search for the optimal model design. Our key insight is to model the design space of possible model designs as a design graph, where the nodes represent design choices, and the edges denote design similarities. FALCON features 1) a task-agnostic module, which performs message passing on the design graph via a Graph Neural Network (GNN), and 2) a task-specific module, which conducts label propagation of the known model performance information on the design graph. Both modules are combined to predict the design performances in the design space, navigating the search direction. We conduct extensive experiments on 27 node and graph classification tasks from various application domains, and an image classification task on the CIFAR-10 dataset. We empirically show that FALCON can efficiently obtain the well-performing designs for each task using only 30 explored nodes. Specifically, FALCON has a comparable time cost with the one-shot approaches while achieving an average improvement of 3.3% compared with the best baselines.


ROLAND: Graph Learning Framework for Dynamic Graphs

August 2022

·

109 Reads

Graph Neural Networks (GNNs) have been successfully applied to many real-world static graphs. However, the success of static graphs has not fully translated to dynamic graphs due to the limitations in model design, evaluation settings, and training strategies. Concretely, existing dynamic GNNs do not incorporate state-of-the-art designs from static GNNs, which limits their performance. Current evaluation settings for dynamic GNNs do not fully reflect the evolving nature of dynamic graphs. Finally, commonly used training methods for dynamic GNNs are not scalable. Here we propose ROLAND, an effective graph representation learning framework for real-world dynamic graphs. At its core, the ROLAND framework can help researchers easily repurpose any static GNN to dynamic graphs. Our insight is to view the node embeddings at different GNN layers as hierarchical node states and then recurrently update them over time. We then introduce a live-update evaluation setting for dynamic graphs that mimics real-world use cases, where GNNs are making predictions and being updated on a rolling basis. Finally, we propose a scalable and efficient training approach for dynamic GNNs via incremental training and meta-learning. We conduct experiments over eight different dynamic graph datasets on future link prediction tasks. Models built using the ROLAND framework achieve on average 62.7% relative mean reciprocal rank (MRR) improvement over state-of-the-art baselines under the standard evaluation settings on three datasets. We find state-of-the-art baselines experience out-of-memory errors for larger datasets, while ROLAND can easily scale to dynamic graphs with 56 million edges. After re-implementing these baselines using the ROLAND training strategy, ROLAND models still achieve on average 15.5% relative MRR improvement over the baselines.



AdaGrid: Adaptive Grid Search for Link Prediction Training Objective

March 2022

·

52 Reads

One of the most important factors that contribute to the success of a machine learning model is a good training objective. Training objective crucially influences the model's performance and generalization capabilities. This paper specifically focuses on graph neural network training objective for link prediction, which has not been explored in the existing literature. Here, the training objective includes, among others, a negative sampling strategy, and various hyperparameters, such as edge message ratio which controls how training edges are used. Commonly, these hyperparameters are fine-tuned by complete grid search, which is very time-consuming and model-dependent. To mitigate these limitations, we propose Adaptive Grid Search (AdaGrid), which dynamically adjusts the edge message ratio during training. It is model agnostic and highly scalable with a fully customizable computational budget. Through extensive experiments, we show that AdaGrid can boost the performance of the models up to 1.9%1.9\% while being nine times more time-efficient than a complete search. Overall, AdaGrid represents an effective automated algorithm for designing machine learning training objectives.


On the Opportunities and Risks of Foundation Models

August 2021

·

11,370 Reads

·

102 Citations

AI is undergoing a paradigm shift with the rise of models (e.g., BERT, DALL-E, GPT-3) that are trained on broad data at scale and are adaptable to a wide range of downstream tasks. We call these models foundation models to underscore their critically central yet incomplete character. This report provides a thorough account of the opportunities and risks of foundation models, ranging from their capabilities (e.g., language, vision, robotics, reasoning, human interaction) and technical principles (e.g., model architectures, training procedures, data, systems, security, evaluation, theory) to their applications (e.g., law, healthcare, education) and societal impact (e.g., inequity, misuse, economic and environmental impact, legal and ethical considerations). Though foundation models are based on conventional deep learning and transfer learning, their scale results in new emergent capabilities, and their effectiveness across so many tasks incentivizes homogenization. Homogenization provides powerful leverage but demands caution, as the defects of the foundation model are inherited by all the adapted models downstream. Despite the impending widespread deployment of foundation models, we currently lack a clear understanding of how they work, when they fail, and what they are even capable of due to their emergent properties. To tackle these questions, we believe much of the critical research on foundation models will require deep interdisciplinary collaboration commensurate with their fundamentally sociotechnical nature.



Identity-aware Graph Neural Networks

May 2021

·

76 Reads

·

343 Citations

Proceedings of the AAAI Conference on Artificial Intelligence

Message passing Graph Neural Networks (GNNs) provide a powerful modeling framework for relational data. However, the expressive power of existing GNNs is upper-bounded by the 1-Weisfeiler-Lehman (1-WL) graph isomorphism test, which means GNNs that are not able to predict node clustering coefficients and shortest path distances, and cannot differentiate between different d-regular graphs. Here we develop a class of message passing GNNs, named Identity-aware Graph Neural Networks (ID-GNNs), with greater expressive power than the 1-WL test. ID-GNN offers a minimal but powerful solution to limitations of existing GNNs. ID-GNN extends existing GNN architectures by inductively considering nodes’ identities during message passing. To embed a given node, ID-GNN first extracts the ego network centered at the node, then conducts rounds of heterogeneous message passing, where different sets of parameters are applied to the center node than to other surrounding nodes in the ego network. We further propose a simplified but faster version of ID-GNN that injects node identity information as augmented node features. Alto- gether, both versions of ID-GNN represent general extensions of message passing GNNs, where experiments show that transforming existing GNNs to ID-GNNs yields on average 40% accuracy improvement on challenging node, edge, and graph property prediction tasks; 3% accuracy improvement on node and graph classification benchmarks; and 15% ROC AUC improvement on real-world link prediction tasks. Additionally, ID-GNNs demonstrate improved or comparable performance over other task-specific graph networks.


Citations (14)


... Graph-Based Representations: Recent advancements in graph-based representations (Lin et al., 2022;Fouladvand et al., 2023) have offered a more versatile and scalable approach to capturing logical relationships within text. Graph neural networks (GNNs) have shown promise in modeling dependencies and hierarchies in various natural language processing tasks (Zhang et al., 2020;Huai et al., 2023;Amador-Domínguez et al., 2023). ...

Reference:

MedLogic-AQA: Enhancing Medical Question Answering with Abstractive Models Focusing on Logical Structures
Graph-based clinical recommender: Predicting specialists procedure orders using graph representation learning
  • Citing Article
  • June 2023

Journal of Biomedical Informatics

... In comparison, subgraph GNNs have been proposed and demonstrated to be more expressive than the 1 WL test (Zhou, Wang, and Zhang 2023) and effective in practice (Bar-Shalom, Bevilacqua, and Maron 2024). Pioneers in this field, (You et al. 2021) and (Zhang and Li 2021) propose ID-GNN and NGNN, respectively. These studies can be regarded as the subgraph GNNs with the node-marking and ego-net policies. ...

Identity-aware Graph Neural Networks
  • Citing Article
  • May 2021

Proceedings of the AAAI Conference on Artificial Intelligence

... We conducted experiments using five real-world dynamic graph datasets, with detailed information provided in Table II, We selected seven baselines for comparison, including both static [1], [8] and dynamic graph neural networks [13], [16], [18], [19], [25]. The primary task is to predict the node labels in the test set at various time slices using the node features and label information from limited time slices in the training set. ...

ROLAND: Graph Learning Framework for Dynamic Graphs
  • Citing Conference Paper
  • August 2022

... In contrast, inductive KG reasoning approaches [77] address this limitation by enabling generalization to KGs with previously unseen entities or relations. Most existing inductive methods [8,29,54,67] employ query-conditional MPNNs to generate "relative" entity embeddings by extracting local structural features from a subgraph induced by the query entity. GraIL [50], for example, extracts an enclosing subgraph between the query entity and each candidate entity, but this approach suffers from high computational costs. ...

Inductive Learning on Commonsense Knowledge Graph Completion
  • Citing Conference Paper
  • July 2021

... Building on advances in self-supervised learning, foundation models represent a recent paradigm shift in AI. 28 These models are trained on massive, diverse datasets, often encompassing internet-scale data. During training, foundation models use self-supervised objectives that enable them to learn representations without explicit labels across massive datasets. ...

On the Opportunities and Risks of Foundation Models

... [19], [41], [42] However, the models were limited to single repeating-unit homopolymers. Furthermore, the lack of distinguishable nodes' embeddings on a computational graph, [43] to embed a long copolymer's structure, impedes the GNN's power in predicting copolymers' properties. As a result, the GNN model cannot differentiate target nodes from the neighboring nodes, failing to provide reasonable predictions. ...

Identity-aware Graph Neural Networks
  • Citing Preprint
  • January 2021

... However, in real-world scenarios, many graphs are extensively large-scale with billions of nodes and edges [20]. Training a single GNN model on such large-scale graphs incurs substantial time and memory costs [21], which leads to much larger overhead for GNAS due to that a large number of candidate architectures need to be trained and evaluated during the search process. Consequently, most GNAS methods fail to handle large-scale graphs. ...

Redundancy-Free Computation for Graph Neural Networks
  • Citing Conference Paper
  • August 2020

... 针对单个图上的组合优化问题 (例如图割、 节点覆盖、 旅行商等), 图建模的方法 较为直观 [11,32,42] ; 若同时存在两个图 (例如图匹配), 则可以采用 "伴随图" 的形式构建融合两个图结 构的大图 [43] ; 对于布尔可满足性问题, 也存在对应的图建模方法. 例如将合取范式等价为二分图 [44] ; 对于一般的矩阵形式 (如线性整数规划), 目前常见的处理方式是将 "Ax b" 等约束条件和线性的问 题参数一起构成一个二分图 [35] . 综上所述, 对多数组合优化问题而言, 均存在其某种合理的图建模方 式. 读者可参阅上述文献, 完成对特定问题的图建模. ...

G2SAT: Learning to Generate SAT Formulas
  • Citing Article
  • December 2019

Advances in Neural Information Processing Systems

... These datasets provide a diverse range of scenarios to evaluate the effectiveness of various methods in predicting drug affinities across different biological and chemical contexts. We further consider four graph-level datasets used in DIR [34]: ❸ SP-Motif [56] is a synthetic dataset, in which the degree of bias can be manually controlled. Each graph consists of a base (with structures from Tree, Ladder, and Wheel) and a motif (with structures from Cycle, House, and Crane). ...

GNNExplainer: Generating Explanations for Graph Neural Networks
  • Citing Article
  • December 2019

Advances in Neural Information Processing Systems