Conference PaperPDF Available

Advanced Hypergraph Mining for Web Applications Using Sphere Neural Networks

Authors:

Abstract and Figures

Web-based applications often involve analyzing complex multi-relational data generated by various domains, including social platforms , bibliographic networks, recommendation systems, and e-commerce platforms. Traditional graph-based methods struggle to model interactions beyond simple pairwise relationships, such as higher-order dependencies and the underlying geometric and structural properties of the data. This paper presents a novel application of hyperspherical deep learning to hypergraphs, integrating geometric hypergraph mining with a Sphere Neural Network (SNN) to model and analyze these intricate relationships effectively. Using real-world datasets, including Reddit, DBLP, MovieLens, and Amazon Co-purchase, our framework embeds hypergraphs into hyperspherical spaces, preserving both relational and geometric properties. Experimental results demonstrate that our method significantly improves performance on tasks such as recommendation, co-purchase prediction, and user behavior analysis, outperforming state-of-the-art techniques. This work highlights the potential of integrating geometric hypergraphs and hyperspherical deep learning to advance the analysis of web-based data.
Content may be subject to copyright.
Advanced Hypergraph Mining for Web Applications Using
Sphere Neural Networks
Zhongtian Sun
zs440@cam.ac.uk
University of Cambridge
Cambridge, UK
University of Oxford
Oxford, UK
University of Kent
Canterbury, UK
Anoushka Harit
University of Cambridge
Cambridge, UK
ah2415@cam.ac.uk
Jongmin Yu
jm.andrew.yu@gmail.com
University of Cambridge
Cambridge, UK
Jingyun Wang
jingyun.wang@durham.ac.uk
Durham University
Durham, UK
Pietro Liò
pl219@cam.ac.uk
University of Cambridge
Cambridge, UK
Abstract
Web-based applications often involve analyzing complex multi-
relational data generated by various domains, including social plat-
forms, bibliographic networks, recommendation systems, and e-
commerce platforms. Traditional graph-based methods struggle
to model interactions beyond simple pairwise relationships, such
as higher-order dependencies and the underlying geometric and
structural properties of the data. This paper presents a novel appli-
cation of hyperspherical deep learning to hypergraphs, integrating
geometric hypergraph mining with a Sphere Neural Network (SNN)
to model and analyze these intricate relationships eectively. Us-
ing real-world datasets, including Reddit, DBLP, MovieLens, and
Amazon Co-purchase, our framework embeds hypergraphs into
hyperspherical spaces, preserving both relational and geometric
properties. Experimental results demonstrate that our method sig-
nicantly improves performance on tasks such as recommendation,
co-purchase prediction, and user behavior analysis, outperforming
state-of-the-art techniques. This work highlights the potential of in-
tegrating geometric hypergraphs and hyperspherical deep learning
to advance the analysis of web-based data.
CCS Concepts
Computing Methodologies
Graph Neural Networks, Hy-
pergraph, Sphere Neural Network.
Keywords
Graph Representation Learning, Recommendation System, Hyper-
graph , Sphere Neural Network
ACM Reference Format:
Zhongtian Sun, Anoushka Harit, Jongmin Yu, Jingyun Wang, and Pietro Liò.
2025. Advanced Hypergraph Mining for Web Applications Using Sphere
This work is licensed under a Creative Commons Attribution Inter-
national 4.0 License.
WWW Companion ’25, Sydney, NSW, Australia
©2025 Copyright held by the owner/author(s).
ACM ISBN 979-8-4007-1331-6/2025/04
https://doi.org/10.1145/3701716.3715577
Neural Networks. In Companion Proceedings of the ACM Web Conference
2025 (WWW Companion ’25), April 28-May 2, 2025, Sydney, NSW, Australia.
ACM, New York, NY, USA, 5 pages. https://doi.org/10.1145/3701716.3715577
1 Introduction
Web-based applications generate vast amounts of complex data,
characterized by multi-relational interactions and higher-order
structures. Examples include user discussions on Reddit [
1
]
1
, aca-
demic collaboration networks on DBLP [
17
]
2
, movie rating patterns
on MovieLens[
6
]
3
, and co-purchase behaviors on Amazon [
9
]
4
.
Traditional graph-based models [
7
,
12
,
14
16
,
18
,
20
] often fail to
capture the nuanced relationships and structural complexity inher-
ent in these datasets, limiting their eectiveness in tasks such as
recommendation, behavior prediction, and anomaly detection.
Hypergraphs, which generalize graphs by allowing hyperedges
to connect multiple nodes, provide a more expressive framework for
modeling such relationships [
2
,
5
,
13
]. However, eectively mining
insights from hypergraphs remains a challenge due to their high-
dimensional nature and lack of appropriate embedding techniques.
To address this, we propose a new application of hyperspherical
deep learning to hypergraph analysis, integrating geometric hyper-
graph mining [
11
] with Sphere Neural Networks (SNNs) [
3
,
19
], a
class of neural networks designed for hyperspherical spaces. By
embedding hypergraph structures into a hyperspherical space [
11
],
our approach captures both geometric and relational properties,
enabling robust analysis of web-based datasets. Our contributions
are threefold:
(1)
We introduce a geometric hypergraph mining framework
tailored to web-based data, leveraging the expressive power
of hyperspherical embeddings [10].
(2)
We develop a novel integration of Sphere Neural Networks
[
3
] with hypergraph learning, enhancing higher-order rela-
tionship modeling and extending SNNs beyond traditional
graphs to more complex structures.
1https://zenodo.org/records/3608135
2https://www.kaggle.com/datasets/dheerajmpai/dblp2023
3https://www.kaggle.com/datasets/grouplens/movielens-20m-dataset
4https://snap.stanford.edu/data/amazon-meta.html
WWW Companion ’25, April 28-May 2, 2025, Sydney, NSW, Australia Zhongtian Sun, Anoushka Harit, Jongmin Yu, Jingyun Wang, & Pietro Liò
(3)
We validate our approach on diverse datasets like Reddit [
1
],
DBLP [
17
], MovieLens [
6
], and Amazon Co-purchase [
9
],
demonstrating its superiority in key tasks such as recom-
mendation and link prediction.
2 Related Work
The analysis of web-based data has seen extensive exploration
through graph-based models. Traditional approaches, such as col-
laborative ltering and matrix factorization [
8
], focus on pairwise
relationships but fail to capture higher-order interactions. Hyper-
graphs, which extend graphs by allowing edges to connect multiple
nodes, oer a richer representation of complex data. Hypergraph-
based models have been applied to tasks like recommendation [4]
and community detection [
23
], yet they often struggle with scala-
bility and the ecient encoding of geometric relationships. Recent
advances in geometric deep learning have introduced neural net-
works for non-Euclidean spaces, such as Sphere Neural Networks
(SNNs) [
3
,
19
], which excel at modeling data in hyperspherical
spaces. However, these methods have primarily been applied to
tasks like image classication and molecular structure prediction,
leaving their potential for web-based hypergraph mining under-
explored.Our work bridges these gaps by combining hypergraph
mining with SNNs[
3
] to enable ecient modeling of multirelational
geometric data in web-based applications. Unlike prior methods,
our framework integrates hyperspherical embeddings to preserve
both geometric and relational properties, oering enhanced per-
formance on datasets such as Reddit[
1
], DBLP [
17
], MovieLens[
6
],
and Amazon Co-purchase[9].
3 Method
Our approach combines geometric hypergraph mining [
11
] with
Sphere Neural Networks (SNNs) [
3
] to model and analyze web-
based data eectively.
3.1 Hypergraph Representation
Let the web-based data be represented as a hypergraph
H=(V,E,
X
)
,
where:
(1) V={𝑣1, 𝑣2, . . . , 𝑣𝑁}is the set of 𝑁nodes.
(2) E={𝑒1, 𝑒2, . . . , 𝑒𝑀}is the set of 𝑀hyperedges.
(3)
X
R𝑁×𝐹
is the node feature matrix, where
𝐹
is the dimen-
sionality of the input features.
Each hyperedge
𝑒𝑘
connects a subset of nodes
V
𝑘 V
. The inci-
dence matrix HR𝑁×𝑀encodes this relationship:
𝐻𝑖,𝑘 =(1if 𝑣𝑖 V
𝑘
0otherwise
3.2 Hyperspherical Embedding
To map the hypergraph to a geometric space, we embed nodes
and hyperedges into a hyperspherical manifold. The embedding
function
𝜙
:
V E S𝑑
projects nodes and hyperedges onto the
𝑑-dimensional unit sphere:
𝜙(𝑣𝑖) S𝑑and 𝜙(𝑒𝑘) S𝑑,where S𝑑={xR𝑑+1:x2=1}.
(1)
The embeddings are optimized to minimize the hyperspherical
distortion while preserving relational and geometric properties.
This is achieved via:
Lembed =
(𝑖, 𝑗 )∈E
𝑤𝑖 𝑗 ·𝜙(𝑣𝑖)𝜙𝑣𝑗2
2+𝛽·
𝑖 V 𝜙(𝑣𝑖) 2
212,
(2)
(𝑖, 𝑗 ) E
is a connected node pair with weight
𝑤𝑖 𝑗 .𝜙 (𝑣𝑖)
is the
hyperspherical embedding,
𝜙(𝑣𝑖)𝜙𝑣𝑗2
2
enforces proximity,
𝛽
regulates constraints, and
𝜙(𝑣𝑖)2
2
1penalizes deviations from
the unit sphere.
3.3 Sphere Neural Network (SNN)
The SNN operates on hyperspherical embeddings to propagate and
aggregate features. It consists of the following layers:
(1)
Spherical Convolution: The spherical convolution layer ag-
gregates features along hyperedges using:
h(𝑙+1)
𝑣=ReLU ©«
𝑒𝑘 N (𝑣)
1
|V
𝑘|
𝑣𝑗 V
𝑘
W(𝑙)h(𝑙)
𝑣𝑗ª®¬,(3)
where h
(𝑙)
𝑣R𝑑
is the feature of node
𝑣
at layer
𝑙, N (𝑣)
is
the set of hyperedges containing
𝑣
, and W
(𝑙)
is the trainable
weight matrix.
(2)
Spherical Attention Mechanism: An attention mechanism
assigns importance scores to hyperedges
𝛼𝑘=
exp a·concat h𝑣𝑖,h𝑒𝑘
Í𝑘exp a·concat h𝑣𝑖,h𝑒𝑘,(4)
where ais a learnable parameter vector. The attention scores
𝛼𝑘are used to weight the hyperedge contributions.
3.4 Task-Specic Objectives
We formulate task-specic loss functions tailored to web-based
applications:
(1)
Recommendation: For link prediction (e.g., co-purchase or
ratings), we maximize the similarity between connected
nodes on the hypersphere:
L𝑟𝑒𝑐 =
(𝑖, 𝑗 )∈E+
log 𝜎𝜙(𝑣𝑖)𝜙(𝑣𝑗)
(𝑖, 𝑗 )∈E
log 1𝜎𝜙(𝑣𝑖)𝜙(𝑣𝑗).
(5)
where
E+
and
E
are the positive and negative edges, respec-
tively, and 𝜎is the sigmoid function.
(2)
Clustering: For community detection, we minimize the intra-
cluster variance while maximizing inter-cluster separation:
Lclust =
𝐶
1
|𝐶|
𝑣𝑖,𝑣𝑗𝐶𝜙(𝑣𝑖)𝜙𝑣𝑗2
2𝜆
𝐶,𝐶
c𝐶c𝐶2
2,
(6)
where c𝐶is the cluster centroid, and 𝜆balances intra-cluster com-
pactness and inter-cluster separation.
3.5 Overall Optimization
The total loss function combines embedding, task-specic, and
regularization terms:
L=Lembed +𝛾Ltask +𝜆Lreg (7)
Advanced Hypergraph Mining for Web Applications Using Sphere Neural Networks WWW Companion ’25, April 28-May 2, 2025, Sydney, NSW, Australia
where
Lembed
preserves relational and geometric properties by
keeping connected nodes close and enforcing the hyperspherical
constraint;
Ltask
optimizes task-specic objectives like link pre-
diction or clustering; and
Lreg
prevents overtting and improves
stability by constraining model parameters. The hyperparameters
𝛾
and
𝜆
(same as used in equation 6) balance task-specic learning
and regularization for optimal generalization.
4 Experiments
We evaluate the proposed HyperSNN framework on four widely
used benchmark hypergraph-structured web datasets, including
Reddit [
1
], DBLP [
17
], MovieLens [
6
], and Amazon Co-Purchase [
9
],
by comparing it with state-of-the-art baselines and conducting an
ablation study to analyze the eectiveness of its components.
4.1 Experiment Setup
4.1.1 Datasets and Baselines. Reddit and DBLP focus on link pre-
diction tasks, while MovieLens and Amazon Co-Purchase target
recommendation and co-purchase prediction tasks. We compare Hy-
perSNN with baselines, including Graph Convolutional Networks
(GCN) [
7
], Graph Attention Networks (GAT) [
18
], Hypergraph GCN
[21], and a Multi-Layer Perceptron (MLP) [22].
4.1.2 Implementation Details. Node features are normalized for
numerical stability and faster convergence. Uniform weights are
assigned to hyperedges for consistency across datasets, and the hy-
pergraph is stored as a sparse incidence matrix to optimize memory
and computation for large-scale data. Models are trained for 100
epochs using the Adam optimizer with a learning rate of 0.01 and
a weight decay of 10
4
for regularization. Experiments are con-
ducted on an NVIDIA GeForce 2080 Ti GPU and evaluated on Area
Under the Curve (AUC), Precision@10, and Mean Reciprocal Rank
(overMRR), with results averaged over ve runs using dierent
random seeds to ensure statistical robustness.
4.2 Results
We evaluate HyperSNN across four datasets (Reddit, DBLP, Movie-
Lens, Amazon Co-Purchase), selecting metrics suited to each task.
For link prediction, AUC measures the ability to distinguish correct
links, while MRR evaluates ranking quality. For recommendation,
P@10 assesses top-10 relevance, and MRR ensures accurate ranking.
These task-specic metrics provide a fair and meaningful evaluation.
The following sections analyze performance, compare baselines,
and present ablation insights.
4.2.1 Link Prediction Task. We present the results of link predic-
tion experiments on Reddit [
1
] and DBLP [
17
] datasets in Table 1.
HyperSNN achieves the highest AUC scores, improving by 6.0% and
5.0% on Reddit and DBLP, respectively, compared to the strongest
baseline. AUC evaluates the model’s ability to rank positive links
higher than negative ones, making it a key metric for link prediction.
Furthermore, HyperSNN exhibits low standard deviation across
runs, highlighting its stability and robustness in capturing complex
hypergraph structures.
Table 1: Link Prediction Results (AUC on Reddit and DBLP)
Model Reddit DBLP
Mean AUC (%) Std Dev (%) Mean AUC (%) Std Dev (%)
GCN 82.0 2.1 80.0 2.3
GAT 84.0 2.0 83.0 2.4
HyperGCN 85.0 1.9 84.0 2.2
MLP 73.0 3.5 72.0 3.6
HyperSNN 91.0 1.7 89.0 1.8
Figure 1: Model Performance for Link Prediction Task
These results highlight HyperSNN’s eectiveness in modeling
higher-order interactions within hypergraphs. Although traditional
models such as GCN [
7
] and GAT [
18
] leverage neighborhood
aggregation, they struggle to capture the geometric and relational
complexity that HyperSNN excels at.
4.2.2 Recommendation and Co-Purchase Prediction Tasks. We eval-
uate HyperSNN on MovieLens[
6
] and Amazon Co-Purchase datasets[
9
]
for recommendation and co-purchase prediction tasks. As shown in
Table 2, HyperSNN outperforms all baselines in both Precision@10,
which measures the relevance of the top 10 recommendations, and
MRR, which evaluates the quality of the rankings. This demon-
strates HyperSNN’s ability to generate and rank relevant recom-
mendations eectively for real-world applications.
Table 2: Recommendation Results (Precision@10 on Movie-
Lens and MRR on Amazon)
Model MovieLens (Precision@10) Amazon (MRR)
Mean (%) Std Dev (%) Mean (%) Std Dev (%)
GCN 76.0 2.5 70.0 2.7
GAT 78.0 2.3 72.0 2.5
HyperGCN 79.0 2.0 73.0 2.2
MLP 65.0 3.4 60.0 3.5
HyperSNN 87.0 1.8 81.0 1.9
WWW Companion ’25, April 28-May 2, 2025, Sydney, NSW, Australia Zhongtian Sun, Anoushka Harit, Jongmin Yu, Jingyun Wang, & Pietro Liò
Figure 2: Model Performance for Recommendation and Co-
Purchase Prediction Tasks
These results demonstrate that the hyperspherical embeddings
learned by HyperSNN provide superior representation power for
user-item and product interactions.
4.3 Ablation Study
We perform an ablation study to assess the contribution of key
components in HyperSNN by systematically removing the atten-
tion mechanism and hyperspherical embeddings. Table 3 and Fig 3
shows that removing the attention mechanism reduces the AUC,
demonstrating its importance in assigning relevance to hyperedges.
Table 3: Ablation Study Results Across All Datasets
Model Variant Reddit DBLP MovieLens Amazon
(AUC) (AUC) (P@10) (MRR)
Full Model 0.91 0.89 0.87 0.81
Without Attention 0.88 0.85 0.84 0.78
Without Hyperspherical 0.82 0.80 0.79 0.73
Without Both 0.79 0.77 0.75 0.70
Figure 3: Ablation Study for HyperSNN
Similarly, replacing hyperspherical embeddings with Euclidean
embeddings results in a signicant drop in performance, highlight-
ing the necessity of geometric representations for capturing re-
lational structures. These results conrm the critical role of both
components in HyperSNN’s success.
5 Discussion
HyperSNN achieves superior performance across all datasets by
leveraging hyperspherical embeddings to capture geometric re-
lationships and an attention mechanism to prioritize critical hy-
peredges. The ablation study highlights the importance of these
components, with signicant performance drops when either is
removed. The spherical convolution layer ensures eective feature
aggregation, preserving the relational and geometric properties of
hypergraphs. The model’s scalability is evident from its ecient
handling of large datasets like Amazon Co-Purchase [
9
], thanks to
sparse representations and batch processing. However, its reliance
on hyperspherical embeddings and attention mechanisms increases
memory usage. This trade-o is oset by its ability to consistently
deliver stable and accurate results, as evidenced by low standard de-
viations across runs. HyperSNN’s success underscores the value of
integrating geometric learning with hypergraph structures, paving
the way for further advancements in this domain.
6 Conclusion
We introduced HyperSNN, a novel framework that integrates hy-
perspherical embeddings and attention mechanisms to eectively
hypergraph modeling. By leveraging geometric representations and
dynamic edge weighting, HyperSNN achieves strong performance
in link prediction, recommendation and co-purchase prediction. Ex-
periments on Reddit, DBLP, MovieLens, and Amazon Co-Purchase
datasets demonstrate its superiority over state-of-the-art baselines,
achieving signicant improvements in AUC, Precision@10, and
MRR.
The ablation study conrms the importance of hyperspherical
embeddings and attention mechanisms, with performance drops
observed when either is removed. HyperSNN also scales eciently
on large datasets using sparse representations, though its memory
requirements remain a limitation. Future work can focus on improv-
ing memory eciency, enhancing robustness to noisy data, and
extending HyperSNN to dynamic hypergraphs, further expanding
its applicability in real-world scenarios.
References
[1]
Jason Baumgartner, Savvas Zannettou, Brian Keegan, Megan Squire, and Jeremy
Blackburn. 2020. The pushshift reddit dataset. In Proceedings of the international
AAAI conference on web and social media, Vol. 14. 830–839.
[2]
C Berge and E Minieka. 1973. Graphs and Hypergraphs. Graphs and Hypergraphs.
[3]
Tiansi Dong, Mateja Jamnik, and Pietro Liò. 2024. Sphere Neural-Networks for
Rational Reasoning. arXiv preprint arXiv:2403.15297 (2024).
[4]
Fuli Feng, Huimin Chen, Xiangnan He, Jie Ding, Maosong Sun, and Tat-Seng
Chua. 2019. Enhancing Stock Movement Prediction with Adversarial Training..
In IJCAI, Vol. 19. 5843–5849.
[5]
Anoushka Harit, Zhongtian Sun, Jongmin Yu, and Noura Al Moubayed. 2024.
Breaking Down Financial News Impact: A Novel AI Approach with Geometric
Hypergraphs. arXiv preprint arXiv:2409.00438 (2024).
[6]
F Maxwell Harper and Joseph A Konstan. 2015. The movielens datasets: History
and context. Acm transactions on interactive intelligent systems (tiis) 5, 4 (2015),
1–19.
[7]
Thomas N Kipf and Max Welling. 2016. Semi-supervised classication with graph
convolutional networks. arXiv preprint arXiv:1609.02907 (2016).
[8]
Yehuda Koren, Robert Bell, and Chris Volinsky. 2009. Matrix factorization tech-
niques for recommender systems. Computer 42, 8 (2009), 30–37.
[9]
Jure Leskovec, Lada A Adamic, and Bernardo A Huberman. 2007. The dynamics
of viral marketing. ACM Transactions on the Web (TWEB) 1, 1 (2007), 5–es.
[10]
Weiyang Liu, Zhen Liu, James M Rehg, and Le Song. 2019. Neural similarity
learning. Advances in Neural Information Processing Systems 32 (2019).
[11]
Danilo Montesi and Giacomo Bergami. [n. d.]. Hypergraph Mining for Social
Networks. ([n. d.]).
Advanced Hypergraph Mining for Web Applications Using Sphere Neural Networks WWW Companion ’25, April 28-May 2, 2025, Sydney, NSW, Australia
[12]
Zhongtian Sun. 2024. Robustness, Heterogeneity and Structure Capturing for
Graph Representation Learning and its Application. P h. D. Dissertation. Durham
University.
[13]
Zhongtian Sun, Anoushka Harit, Alexandra I Cristea, Jingyun Wang, and Pietro
Lio. 2023. Money: Ensemble learning for stock price movement prediction via
a convolutional network with adversarial hypergraph model. AI Open 4 (2023),
165–174.
[14]
Zhongtian Sun, Anoushka Harit, Alexandra I Cristea, Jingyun Wang, and Pietro
Lio. 2023. A Rewiring Contrastive Patch PerformerMixer Framework for Graph
Representation Learning. In 2023 IEEE International Conference on Big Data (Big-
Data). IEEE Computer Society, 5930–5939.
[15]
Zhongtian Sun, Anoushka Harit, Alexandra I Cristea, Jialin Yu, Lei Shi, and Noura
Al Moubayed. 2022. Contrastive learning with heterogeneous graph attention
networks on short text classication. In 2022 International Joint Conference on
Neural Networks (IJCNN). IEEE, 1–6.
[16]
Zhongtian Sun, Anoushka Harit, Jialin Yu, Alexandra I Cristea, and Noura
Al Moubayed. 2021. A generative bayesian graph attention network for semi-
supervised classication on scarce data. In 2021 International Joint Conference on
Neural Networks (IJCNN). IEEE, 1–7.
[17]
The DBLP Team. 2019. DBLP Computer Science Bibliography: Monthly Snapshot
Release of November 2019. https://dblp.org/xml/release/dblp- 2019-11-01.xml.gz.
Accessed: 2019-11-01.
[18]
Petar Velickovic, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro
Lio, Yoshua Bengio, et al
.
2017. Graph attention networks. stat 1050, 20 (2017),
10–48550.
[19]
Richard C Wilson, Edwin R Hancock, Elżbieta Pekalska, and Robert PW Duin.
2014. Spherical and hyperbolic embeddings of data. IEEE transactions on pattern
analysis and machine intelligence 36, 11 (2014), 2255–2269.
[20]
Adam Wynn, Jingyun Wang, Zhongtian Sun, and Atsushi Shimada. 2024.
Analysing Learner Behaviour in an Ontology-Based E-learning System: A Graph
Neural Network Approach. (2024).
[21]
Naganand Yadati, Madhav Nimishakavi, Prateek Yadav, Vikram Nitin, Anand
Louis, and Partha Talukdar. 2019. Hypergcn: A new method for training graph
convolutional networks on hypergraphs. Advances in neural information process-
ing systems 32 (2019).
[22]
Shuai Zhang, Lina Yao, Aixin Sun, and Yi Tay. 2019. Deep learning based recom-
mender system: A survey and new perspectives. ACM computing surveys (CSUR)
52, 1 (2019), 1–38.
[23]
Dengyong Zhou, Jiayuan Huang, and Bernhard Schölkopf. 2006. Learning with
hypergraphs: Clustering, classication, and embedding. Advances in neural
information processing systems 19 (2006).
... However, these models typically rely on predefined pairwise residue interactions, which fail to capture the complexity of protein folding. Hypergraph neural networks [11], [24] extend traditional GNNs [20] by modeling multi-residue interactions simultaneously, making them more suitable for protein structure modeling. TextFold builds on this by introducing a geometric hypergraph representation that explicitly encodes higher-order spatial dependencies, enabling improved predictions in complex structural environments. ...
Preprint
Accurate protein structure prediction remains a key challenge in computational biology, especially for novel folds and complex interactions. We present TextFold, a framework that integrates geometric hypergraphs with multimodal textual context to enhance protein folding predictions. By modeling higher-order spatial relationships and incorporating scientific insights from PubMed and PubTator Central scientific database, TextFold improves accuracy and interpretability through feature attribution analysis.Evaluated on PDB and AlphaFold Protein Structure Database, TextFold achieves a TM-score of 0.81 and an RMSD of 2.1Å for homologous and 3.1Å for low-homology proteins, outperforming DeepFold, AlphaFold, and RoseTTAFold in low-homology settings. Ablation studies demonstrate the impact of textual embeddings on prediction refinement.By integrating geometric modeling with domain knowledge, TextFold advances protein structure prediction, offering a valuable tool for drug discovery and functional genomics.
... Recent advances in Graph Neural Networks (GNNs) and Large Language Models (LLMs) present an opportunity to bridge this gap. GNNs effectively model complex relationships [19,17,18,9,22,20], but typically lack integrated feedback mechanisms. Conversely, while LLMs have exhibited considerable potential in generating educational content [11], their efficacy remains limited in predictive tasks and pedagogical interventions within educational contexts. ...
Conference Paper
Most knowledge tracing systems prioritise performance prediction over the delivery of personalised, actionable feedback. This paper proposes SPAR-GNN (Selective Pattern-Aware Reasoning with Graph Neural Networks), a novel framework that combines heterogeneous GNN-based modelling with selectively triggered feedback from a Large Language Model(LLM). SPAR-GNN represents students, problems , skills and virtual pattern nodes that encode latent behavioural traits such as frustration and hint overuse within a unified graph structure. A customised Heterogeneous Graph Layer learns student representations , while the LLM is activated only for at-risk learners identified through behavioural and performance indicators. This reduces computational overhead while enhancing pedagogical relevance. SPAR-GNN consistently outperforms strong baselines across multiple metrics, demonstrating both predictive accuracy and robustness. Observational analysis also reveals interpretable links between behaviours and learning progression , supporting fairness-aware interventions.
Conference Paper
Full-text available
Despite the prevalence of e-learning systems, there is a lack of support for learners to identify and compare new knowledge with existing cognitive structures. Therefore, an ontology-based visualization support system was previously introduced which offers two modes: cache-cache, where relations are initially hidden and the learners are encouraged to create those relations, and receptive, where learners can view expert-generated topic maps. In this study, we aim to analyse learner behaviour by representing user behaviour as graphs and utilising a heterogeneous graph convolutional network. Two graphs are constructed for each student to capture behaviour before and after system use. Results indicate significant differences in mean embeddings between learners in receptive and cache-cache modes. Further analysis, considering pre-test performance, shows no significant differences in the receptive and cache-cache groups but highlights a considerably smaller mean for high prior performers in the cache-cache group.
Article
Full-text available
With the ever-growing volume of online information, recommender systems have been an effective strategy to overcome such information overload. The utility of recommender systems cannot be overstated, given its widespread adoption in many web applications, along with its potential impact to ameliorate many problems related to over-choice. In recent years, deep learning has garnered considerable interest in many research fields such as computer vision and natural language processing, owing not only to stellar performance but also the attractive property of learning feature representations from scratch. The influence of deep learning is also pervasive, recently demonstrating its effectiveness when applied to information retrieval and recommender systems research. Evidently, the field of deep learning in recommender system is flourishing. This article aims to provide a comprehensive review of recent research efforts on deep learning based recommender systems. More concretely, we provide and devise a taxonomy of deep learning based recommendation models, along with providing a comprehensive summary of the state-of-the-art. Finally, we expand on current trends and provide new perspectives pertaining to this new exciting development of the field.
Conference Paper
This paper contributes a new machine learning solution for stock movement prediction, which aims to predict whether the price of a stock will be up or down in the near future. The key novelty is that we propose to employ adversarial training to improve the generalization of a neural network prediction model. The rationality of adversarial training here is that the input features to stock prediction are typically based on stock price, which is essentially a stochastic variable and continuously changed with time by nature. As such, normal training with static price-based features (e.g. the close price) can easily overfit the data, being insufficient to obtain reliable models. To address this problem, we propose to add perturbations to simulate the stochasticity of price variable, and train the model to work well under small yet intentional perturbations. Extensive experiments on two real-world stock data show that our method outperforms the state-of-the-art solution [Xu and Cohen, 2018] with 3.11% relative improvements on average w.r.t. accuracy, validating the usefulness of adversarial training for stock prediction task.
Article
We present graph attention networks (GATs), novel neural network architectures that operate on graph-structured data, leveraging masked self-attentional layers to address the shortcomings of prior methods based on graph convolutions or their approximations. By stacking layers in which nodes are able to attend over their neighborhoods' features, we enable (implicitly) specifying different weights to different nodes in a neighborhood, without requiring any kind of costly matrix operation (such as inversion) or depending on knowing the graph structure upfront. In this way, we address several key challenges of spectral-based graph neural networks simultaneously, and make our model readily applicable to inductive as well as transductive problems. Our GAT models have achieved state-of-the-art results across three established transductive and inductive graph benchmarks: the Cora and Citeseer citation network datasets, as well as a protein-protein interaction dataset (wherein test graphs are entirely unseen during training).
Article
Many computer vision and pattern recognition problems may be posed as the analysis of a set of dissimilarities between objects. For many types of data, these dissimilarities are not euclidean (i.e., they do not represent the distances between points in a euclidean space), and therefore cannot be isometrically embedded in a euclidean space. Examples include shape-dissimilarities, graph distances and mesh geodesic distances. In this paper, we provide a means of embedding such non-euclidean data onto surfaces of constant curvature. We aim to embed the data on a space whose radius of curvature is determined by the dissimilarity data. The space can be either of positive curvature (spherical) or of negative curvature (hyperbolic). We give an efficient method for solving the spherical and hyperbolic embedding problems on symmetric dissimilarity data. Our approach gives the radius of curvature and a method for approximating the objects as points on a hyperspherical manifold without optimisation. For objects which do not reside exactly on the manifold, we develop a optimisation-based procedure for approximate embedding on a hyperspherical manifold. We use the exponential map between the manifold and its local tangent space to solve the optimisation problem locally in the euclidean tangent space. This process is efficient enough to allow us to embed data sets of several thousand objects. We apply our method to a variety of data including time warping functions, shape similarities, graph similarity and gesture similarity data. In each case the embedding maintains the local structure of the data while placing the points in a metric space.