Wenbin Zhang’s research while affiliated with Kunming University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (118)


Workflow of our proposed method
Number of labels in each indicator
Flowchart of BERT-based pre-processing method for textual data
The architecture of our proposed model
Attention mechanism analysis
Forecasting leading economic indicators in the US from financial news using multi-task learning
  • Article
  • Publisher preview available

May 2025

·

2 Reads

Soft Computing

Josh Jia-Ching Ying

·

Chia-Chen Liu

·

Vincent S. Tseng

·

[...]

·

Ji Zhang

Leading economic indicators are crucial statistics that are based on the economy and have a significant impact on various factors such as policies, stock market trends, etc. However, due to the long release time interval of these indicators, it is challenging to accurately predict their trends in advance. In this paper, we propose an effective framework for forecasting leading economic indicators in the United States by utilizing leading indicators to predict each other and learn their mutual correlation through an attention mechanism. To achieve this, we present a hierarchical, multi-task learning approach that uses textual data to make predictions for four essential leading economic indicators. We first group news articles by topic and then summarize and extract information using a time-series method and attention. Finally, we concatenate news information with historical data to forecast the leading economic indicators. Our experimental results demonstrate that our approach successfully forecasts leading economic indicators, and we are able to predict four indicators one month beforehand. This framework has significant potential for use in the finance industry and can be used to inform important decisions related to investment strategies, policymaking, and other financial aspects.

View access options

An OCRNet-Based Method for Assessing Apple Watercore in Cold and Cool Regions of Yunnan Province

May 2025

·

1 Read

The content of the watercore in apples plays a decisive role in their taste and selling price, but there is a lack of methods to accurately assess it. Therefore, this paper proposes an OCRNet-based method for apple watercore content evaluation. A total of 720 watercores of apples from Mengzi, Lijiang, and Zhaotong City in Yunnan Province were used as experimental samples. An appropriate watercore extraction model was selected based on different evaluation indicators. The watercore feature images extracted using the optimal model were stacked, and the watercore content of apples in different regions was evaluated by calculating the fitted area of the stacked watercore region. The results show that the OCRNet model is optimal in all evaluation metrics when facing different datasets. The error of OCRNet is also minimized when extracting overexposed as well as underexposed images with 0.15% and 0.38%, respectively, and it can be used to extract the characteristics of the apple watercore. The evaluation result of the watercore content of apples in different regions is that Lijiang apples have the highest watercore content, followed by Mengzi apples, and Zhaotong apples have the least watercore content.



Toward Adaptive Large Language Models Structured Pruning via Hybrid-grained Weight Importance Assessment

April 2025

·

8 Reads

·

1 Citation

Proceedings of the AAAI Conference on Artificial Intelligence

Structured pruning for large language models (LLMs) has garnered significant academic interest due to its ability to efficiently compress and accelerate LLMs by eliminating redundant weight groups at a coarse-grained granularity. Current structured pruning methods for LLMs typically depend on a singular granularity for assessing weight importance, resulting in notable performance degradation in downstream tasks. Intriguingly, our empirical investigations reveal that utilizing unstructured pruning, which achieves better performance retention by pruning weights at a finer granularity, \emph{i.e.}, individual weights, yields significantly varied sparse LLM structures when juxtaposed to structured pruning. This suggests that evaluating both holistic and individual assessments for weight importance are essential for LLM pruning. Building on this insight, we introduce the Hybrid-grained Weight Importance Assessment (HyWIA), a novel method that merges fine-grained and coarse-grained evaluations of weight importance for the pruning of LLMs. Leveraging an attention mechanism, HyWIA adaptively determines the optimal blend of granularity in weight importance assessments in an end-to-end pruning manner. Extensive experiments on LLaMA-V1/V2, Vicuna, Baichuan, and Bloom across various benchmarks demonstrate the effectiveness of HyWIA in pruning LLMs. For example, HyWIA surpasses the cutting-edge LLM-Pruner by an average margin of 2.82% in accuracy across seven downstream tasks when pruning LLaMA-7B by 50%.


Fair Graph U-Net: A Fair Graph Learning Framework Integrating Group and Individual Awareness

April 2025

·

6 Reads

·

1 Citation

Proceedings of the AAAI Conference on Artificial Intelligence

Learning high-level representations for graphs is crucial for tasks like node classification, where graph pooling aggregates node features to provide a holistic view that enhances predictive performance. Despite numerous methods that have been proposed in this promising and rapidly developing research field, most efforts to generalize the pooling operation to graphs are primarily performance-driven, with fairness issues largely overlooked: i) the process of graph pooling could exacerbate disparities in distribution among various subgroups; ii) the resultant graph structure augmentation may inadvertently strengthen intra-group connectivity, leading to unintended inter-group isolation. To this end, this paper extends the initial effort on fair graph pooling to the development of fair graph neural networks, while also providing a unified framework to collectively address group and individual graph fairness. Our experimental evaluations on multiple datasets demonstrate that the proposed method not only outperforms state-of-the-art baselines in terms of fairness but also achieves comparable predictive performance.



Digital Forensics in the Age of Large Language Models

April 2025

·

65 Reads

Digital forensics plays a pivotal role in modern investigative processes, utilizing specialized methods to systematically collect, analyze, and interpret digital evidence for judicial proceedings. However, traditional digital forensic techniques are primarily based on manual labor-intensive processes, which become increasingly insufficient with the rapid growth and complexity of digital data. To this end, Large Language Models (LLMs) have emerged as powerful tools capable of automating and enhancing various digital forensic tasks, significantly transforming the field. Despite the strides made, general practitioners and forensic experts often lack a comprehensive understanding of the capabilities, principles, and limitations of LLM, which limits the full potential of LLM in forensic applications. To fill this gap, this paper aims to provide an accessible and systematic overview of how LLM has revolutionized the digital forensics approach. Specifically, it takes a look at the basic concepts of digital forensics, as well as the evolution of LLM, and emphasizes the superior capabilities of LLM. To connect theory and practice, relevant examples and real-world scenarios are discussed. We also critically analyze the current limitations of applying LLMs to digital forensics, including issues related to illusion, interpretability, bias, and ethical considerations. In addition, this paper outlines the prospects for future research, highlighting the need for effective use of LLMs for transparency, accountability, and robust standardization in the forensic process.


Apple Watercore Grade Classification Method Based on ConvNeXt and Visible/Near-Infrared Spectroscopy

March 2025

·

11 Reads

To address the issues of insufficient rigor in existing methods for quantifying apple watercore severity and the complexity and low accuracy of traditional classification models, this study proposes a method for watercore quantification and a classification model based on a deep convolutional neural network. Initially, visible/near-infrared transmission spectral data of apple samples were collected. The apples were then sliced into 4.5 mm thick sections using a specialized tool, and image data of each slice were captured. Using BiSeNet and RIFE algorithms, a three-dimensional model of the watercore regions was constructed from the apple slices to calculate the watercore severity, which was subsequently categorized into five distinct levels. Next, methods such as the Gramian Angular Summation Field (GASF), Gram Angular Difference Field (GADF), and Markov Transition Field (MTF) were applied to transform the one-dimensional spectral data into two-dimensional images. These images served as input for training and prediction using the ConvNeXt deep convolutional neural network. The results indicated that the GADF method yielded the best performance, achieving a test set accuracy of 98.73%. Furthermore, the study contrasted the classification and prediction of watercore apples using traditional methods with the existing quantification approaches for watercore levels. The comparative results demonstrated that the proposed GADF-ConvNeXt model is more straightforward and efficient, achieving superior performance in classifying watercore grades. Furthermore, the newly proposed quantification method for watercore levels proved to be more effective.


(a) Forest plot of odds ratios with 95% CI (log scale). (b) Forest plot of odds ratios with 95% CI (log scale).
Mean ROC curve for the three classification models predicting emergency department hospitalization admission.
Demographic and clinical characteristics of emergency department patients categorized by hospital admission status.
Machine learning-driven prediction of hospital admissions using gradient boosting and GPT-2

March 2025

·

10 Reads

Background Accurately predicting hospital admissions from the emergency department (ED) is essential for improving patient care and resource allocation. This study aimed to predict hospital admissions by integrating both structured clinical data and unstructured text data using machine learning models. Methods Data were obtained from the 2021 National Hospital Ambulatory Medical Care Survey—Emergency Department (NHAMCS-ED), including adult patients aged 18 years and older. Structured data included demographics, visit characteristics, vital signs, and medical history, while unstructured data consisted of free-text chief complaints and injury descriptions. A Gradient Boosting Classifier (GBC) was applied to structured data, while a fine-tuned GPT-2 model processed the unstructured text. A combined model was created by averaging the outputs of both models. Model performance was evaluated using 5-fold cross-validation, assessing accuracy, precision, sensitivity, specificity, and area under the receiver operating characteristic curve (AUC-ROC). Results Among the 13,115 patients, 2264 (17.3%) were admitted to the hospital. The combined model outperformed the individual structured and unstructured models, achieving an accuracy of 75.8%, precision of 39.5%, sensitivity of 75.8%, and specificity of 75.8%. In comparison, the structured data model achieved 73.8% accuracy, while the unstructured model reached 64.6%. The combined model had the highest AUC, indicating superior performance. Conclusions Combining structured and unstructured data using machine learning significantly improves the prediction of hospital admissions from the ED. This integrated approach can enhance decision-making and optimize ED operations.


Attention Mechanism-Based Cognition-Level Scene Understanding

March 2025

·

2 Reads

Given a question–image input, a visual commonsense reasoning (VCR) model predicts an answer with a corresponding rationale, which requires inference abilities based on real-world knowledge. The VCR task, which calls for exploiting multi-source information as well as learning different levels of understanding and extensive commonsense knowledge, is a cognition-level scene understanding challenge. The VCR task has aroused researchers’ interests due to its wide range of applications, including visual question answering, automated vehicle systems, and clinical decision support. Previous approaches to solving the VCR task have generally relied on pre-training or exploiting memory with long-term dependency relationship-encoded models. However, these approaches suffer from a lack of generalizability and a loss of information in long sequences. In this work, we propose a parallel attention-based cognitive VCR network, termed PAVCR, which fuses visual–textual information efficiently and encodes semantic information in parallel to enable the model to capture rich information for cognition-level inference. Extensive experiments show that the proposed model yields significant improvements over existing methods on the benchmark VCR dataset. Moreover, the proposed model provides an intuitive interpretation of visual commonsense reasoning.


Citations (58)


... To address this problem, we propose a method that uses an optimized scaling factor (OpS) α/ √ r for fine-tuning both uncompressed and pruned [28], [29], [30], [31] [32], [33] LLMs. While rsLoRA [34] employs a similar scaling factor to study the impact of the scaling factor on the learning process, our motivation, theoretical derivation, and experimental design are independent. ...

Reference:

RoRA: Efficient Fine-Tuning of LLM with Reliability Optimization for Rank Adaptation
Toward Adaptive Large Language Models Structured Pruning via Hybrid-grained Weight Importance Assessment

Proceedings of the AAAI Conference on Artificial Intelligence

... The notion of graph individual fairness, which requires that similar individuals be treated similarly, is appealing, it is, however, difficult to operationalize in practice (Wang, Chu, et al. 2025). For instance, it is unclear what the correct fair way to select features to match is (e.g., whether to match based on a node's attributes, and/or its local/global graph structure). ...

Fair Graph U-Net: A Fair Graph Learning Framework Integrating Group and Individual Awareness
  • Citing Article
  • April 2025

Proceedings of the AAAI Conference on Artificial Intelligence

... RoRA (Liu, Kong et al. 2025) proposes Rank-adaptive Reliability Optimization (RoRA), which optimizes LoRA's scaling factor by replacing α/r with α/ √ r, ensuring improved performance. AdaPTwin (Biju, Sriram, and Pilanci 2024) compresses pairs of weight matrices that are dependent on products within the transformer attention layer simultaneously. ...

RoRA: Efficient Fine-Tuning of LLM with Reliability Optimization for Rank Adaptation
  • Citing Conference Paper
  • April 2025

... The problem of fairness [7]- [14] in machine learning models used has shown to be a issue. Reviewing previous research shows that 67% of papers selected that covered applications of machine learning in the medical domain showed that a racial bias existed in their model [15]. ...

FG-SMOTE: Towards Fair Node Classification with Graph Neural Network
  • Citing Article
  • January 2025

ACM SIGKDD Explorations Newsletter

... Moreover, unlike independent and identically distributed (IID) data, where each instance is treated as independent, graph-based data is inherently non-IID, as nodes are interconnected and impact each other's predictions through message-passing mechanisms in GNNs [66]. Consequently, fairness testing methods developed for IID settings [55,62,63] are ill-suited for graph-structured data, highlighting the need for GNN-specific fairness testing approaches. ...

Fairness amidst non‐IID graph data: A literature review

... In contrast, the SCAI-BIO/BioGottBERT model was further fine-tuned via a small-annotated dataset. Owing to its BERT-based architecture, it captures contextual dependencies and semantic nuances more effectively than traditional NLP methods such as bag-of-words or TF-IDF (term frequency-inverse document frequency) (Zhang et al. 2024). ...

Integrating structured and unstructured data for predicting emergency severity: an association and predictive study using transformer-based natural language processing models

BMC Medical Informatics and Decision Making

... Efficient learning for LLMs. The goal of efficient learning (Zhan et al. 2021;Zhan, Wu et al. 2024;Liu et al. 2024aLiu et al. ,b, 2021Liu et al. , 2022Liu et al. , 2023aLi, Kong et al. 2020) is to achieve better results with fewer resources. SpQR (Dettmers et al. 2023b) employed a method involving the identification and isolation of outlier weights. ...

TSLA: A Task-Specific Learning Adaptation for Semantic Segmentation on Autonomous Vehicles Platform
  • Citing Article
  • January 2024

IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems

... Digital forensics is a critical component in modern investigative and judicial processes, which involve the systematic collection, analysis, and preservation of digital evidence from electronic devices and online activities [1][2][3]. Its primary objective is to uncover factual information related to cybercrimes, fraud, unauthorized access, and other illicit activities [4]. ...

Transforming Digital Forensics with Large Language Models: Unlocking Automation, Insights, and Justice
  • Citing Conference Paper
  • October 2024

... To address these substantial challenges in digital forensics, recent advances in artificial intelligence offer promising solutions. Notably, large language models (LLMs), such as the generative pre-trained transformer (GPT) series of models and the Gemini series, have emerged as powerful tools with the potential to transform digital forensic practices [23,24,30,45]. These advanced AI models are designed to understand, interpret, generate and analyze human language with unprecedented accuracy. ...

Fairness in Large Language Models in Three Hours
  • Citing Conference Paper
  • October 2024

... To address this issue, a number of approaches have been proposed to quantify and mitigate discrimination amidst non-I.I.D. graph data (Wang, Ulloa, et al. 2024;Gupta and Dukkipati 2021;Ma et al. 2022), although graph fairness has largely remained nascent. This paper presents the first thorough study of fairness amidst non-I.I.D. graph data for researchers and practitioners to have a better understanding of current landscape of graph fairness while complementing the existing fairness surveys that largely focus on traditional IID data and fair graph classification. ...

Individual Fairness with Group Constraints in Graph Neural Networks
  • Citing Chapter
  • October 2024