Fan Zhang’s research while affiliated with Shenzhen Bay Laboratory and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (6)


Unified Enhancement of the Generalization and Robustness of Language Models via Bi-Stage Optimization
  • Article

March 2025

·

17 Reads

IACAPAP ArXiv

Yudao Sun

·

Juan Yin

·

Juan Zhao

·

[...]

·

Hongji Chen


Fig. 1. Bi-stage optimization framework. This figure illustrates the employment of adaptive dropout regularization during the forward propagation stage and selective update during the backward propagation stage in the UEGR method, aimed at enhancing language model generalization and robustness. The solid lines denote the forward propagation process, whereas the dashed lines indicate the backward propagation process. Blue lines signify dropped connections within the language model, and red lines signify retained connections.
Fig. 2. A qualitative analysis of UEGR is provided as follows: (a) Examining the similarity between the original pre-trained model and the fine-tuned model, and (b) Assessing the model's robustness under varying levels of adversarial perturbation, along with the improvements achieved through UEGR.
Unified Enhancement of the Generalization and Robustness of Language Models via Bi-Stage Optimization
  • Preprint
  • File available

March 2025

·

23 Reads

Neural network language models (LMs) are confronted with significant challenges in generalization and robustness. Currently, many studies focus on improving either generalization or robustness in isolation, without methods addressing both aspects simultaneously, which presents a significant challenge in developing LMs that are both robust and generalized. In this paper, we propose a bi-stage optimization framework to uniformly enhance both the generalization and robustness of LMs, termed UEGR. Specifically, during the forward propagation stage, we enrich the output probability distributions of adversarial samples by adaptive dropout to generate diverse sub models, and incorporate JS divergence and adversarial losses of these output distributions to reinforce output stability. During backward propagation stage, we compute parameter saliency scores and selectively update only the most critical parameters to minimize unnecessary deviations and consolidate the model's resilience. Theoretical analysis shows that our framework includes gradient regularization to limit the model's sensitivity to input perturbations and selective parameter updates to flatten the loss landscape, thus improving both generalization and robustness. The experimental results show that our method significantly improves the generalization and robustness of LMs compared to other existing methods across 13 publicly available language datasets, achieving state-of-the-art (SOTA) performance.

Download


Reliable Task Offloading in Sustainable Edge Computing With Imperfect Channel State Information

December 2024

·

4 Reads

·

1 Citation

IEEE Transactions on Network and Service Management

As a promising paradigm, edge computing enhances service provisioning by offloading tasks to powerful servers at the network edge. Meanwhile, Non-Orthogonal Multiple Access (NOMA) and renewable energy sources are increasingly adopted for spectral efficiency and carbon footprint reduction. However, these new techniques inevitably introduce reliability risks to the edge system generally because of i) imperfect Channel State Information (CSI), which can misguide offloading decisions and cause transmission outages, and ii) unstable renewable energy supply, which complicates device availability. To tackle these issues, we first establish a system model that measures service reliability based on probabilistic principles for the NOMA-based edge system. As a solution, a Reliable Offloading method with Multi-Agent deep reinforcement learning (ROMA) is proposed. In ROMA, we first reformulate the reliability-critical constraint into an long-term optimization problem via Lyapunov optimization. We discretize the hybrid action space and convert the resource allocation on edge servers into a 0-1 knapsack problem. The optimization problem is then formulated as a Partially Observable Markov Decision Process (POMDP) and addressed by multi-agent proximal policy optimization (PPO). Experimental evaluations demonstrate the superiority of ROMA over existing methods in reducing grid energy costs and enhancing system reliability, achieving Pareto-optimal performance under various settings.