Wenyi Zhang’s research while affiliated with University of Electronic Science and Technology of China and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (210)


Soft Classification in a Composite Source Model
  • Article

June 2025

·

2 Reads

Yuefeng Cao

·

Jiakun Liu

·

Wenyi Zhang

A composite source model consists of an intrinsic state and an extrinsic observation. The fundamental performance limit of reproducing the intrinsic state is characterized by the indirect rate–distortion function. In a remote classification application, a source encoder encodes the extrinsic observation (e.g., image) into bits, and a source decoder plays the role of a classifier that reproduces the intrinsic state (e.g., label of image). In this work, we characterize the general structure of the optimal transition probability distribution, achieving the indirect rate–distortion function. This optimal solution can be interpreted as a “soft classifier”, which generalizes the conventionally adopted “classify-then-compress” scheme. We then apply the soft classification to aid the lossy compression of the extrinsic observation of a composite source. This leads to a coding scheme that exploits the soft classifier to guide reproduction, outperforming existing coding schemes without classification or with hard classification.


Fig. 1. Transceiver architecture.
Fig. 12. Numerical results, asymptotics, and approximations of L * .
An Information-Theoretic Framework for Receiver Quantization in Communication
  • Preprint
  • File available

May 2025

·

2 Reads

We investigate information-theoretic limits and design of communication under receiver quantization. Unlike most existing studies, this work is more focused on the impact of resolution reduction from high to low. We consider a standard transceiver architecture, which includes i.i.d. complex Gaussian codebook at the transmitter, and a symmetric quantizer cascaded with a nearest neighbor decoder at the receiver. Employing the generalized mutual information (GMI), an achievable rate under general quantization rules is obtained in an analytical form, which shows that the rate loss due to quantization is log(1+γSNR)\log\left(1+\gamma\mathsf{SNR}\right), where γ\gamma is determined by thresholds and levels of the quantizer. Based on this result, the performance under uniform receiver quantization is analyzed comprehensively. We show that the front-end gain control, which determines the loading factor of quantization, has an increasing impact on performance as the resolution decreases. In particular, we prove that the unique loading factor that minimizes the MSE also maximizes the GMI, and the corresponding irreducible rate loss is given by log(1+mmseSNR)\log\left(1+\mathsf {mmse}\cdot\mathsf{SNR}\right), where mmse is the minimum MSE normalized by the variance of quantizer input, and is equal to the minimum of γ\gamma. A geometrical interpretation for the optimal uniform quantization at the receiver is further established. Moreover, by asymptotic analysis, we characterize the impact of biased gain control, including how small rate losses decay to zero and achievable rate approximations under large bias. From asymptotic expressions of the optimal loading factor and mmse, approximations and several per-bit rules for performance are also provided. Finally we discuss more types of receiver quantization and show that the consistency between achievable rate maximization and MSE minimization does not hold in general.

Download

AB-Cache: Training-Free Acceleration of Diffusion Models via Adams-Bashforth Cached Feature Reuse

April 2025

·

8 Reads

Zichao Yu

·

Zhen Zou

·

Guojiang Shao

·

[...]

·

Wenyi Zhang

Diffusion models have demonstrated remarkable success in generative tasks, yet their iterative denoising process results in slow inference, limiting their practicality. While existing acceleration methods exploit the well-known U-shaped similarity pattern between adjacent steps through caching mechanisms, they lack theoretical foundation and rely on simplistic computation reuse, often leading to performance degradation. In this work, we provide a theoretical understanding by analyzing the denoising process through the second-order Adams-Bashforth method, revealing a linear relationship between the outputs of consecutive steps. This analysis explains why the outputs of adjacent steps exhibit a U-shaped pattern. Furthermore, extending Adams-Bashforth method to higher order, we propose a novel caching-based acceleration approach for diffusion models, instead of directly reusing cached results, with a truncation error bound of only O(hk)O(h^k) where h is the step size. Extensive validation across diverse image and video diffusion models (including HunyuanVideo and FLUX.1-dev) with various schedulers demonstrates our method's effectiveness in achieving nearly 3×3\times speedup while maintaining original performance levels, offering a practical real-time solution without compromising generation quality.



Fig. 1. Achievable rate (9) with b-bits uniform output quantization.
Fig. 5. Numerical results and asymptotics of L * .
A High-Resolution Analysis of Receiver Quantization in Communication

January 2025

·

8 Reads

We investigate performance limits and design of communication in the presence of uniform output quantization with moderate to high resolution. Under independent and identically distributed (i.i.d.) complex Gaussian codebook and nearest neighbor decoding rule, an achievable rate is derived in an analytical form by the generalized mutual information (GMI). The gain control before quantization is shown to be increasingly important as the resolution decreases, due to the fact that the loading factor (normalized one-sided quantization range) has increasing impact on performance. The impact of imperfect gain control in the high-resolution regime is characterized by two asymptotic results: 1) the rate loss due to overload distortion decays exponentially as the loading factor increases, and 2) the rate loss due to granular distortion decays quadratically as the step size vanishes. For a 2K-level uniform quantizer, we prove that the optimal loading factor that maximizes the achievable rate scales like 2ln2K2\sqrt{\ln 2K} as the resolution increases. An asymptotically tight estimate of the optimal loading factor is further given, which is also highly accurate for finite resolutions.


Generalized Nearest Neighbor Decoding: General Input Constellation and a Case Study of Interference Suppression

January 2025

IEEE Transactions on Communications

In this work, generalized nearest neighbor decoding (GNND), a recently proposed receiver architecture, is studied for channels under general input constellations, and multiuser uplink interference suppression is employed as a case study for demonstrating its potential. In essence, GNND generalizes the well-known nearest neighbor decoding, by introducing a symbol-level memoryless processing step, which can be rendered seamlessly compatible with Gaussian channel-based decoders. First, criteria of the optimal GNND are derived for general input constellations, expressed in the form of conditional moments matching, thereby generalizing the prior work which has been confined to Gaussian input. Then, the optimal GNND is applied to the use case of multiuser uplink, for which the optimal GNND is shown to be capable of achieving information rates nearly identical to the channel mutual information. By contrast, the commonly used channel linearization (CL) approach incurs a noticeable rate loss. A coded modulation scheme is subsequently developed, aiming at implementing GNND using off-the-shelf channel codes, without requiring iterative message passing between demodulator and decoder. Through numerical experiments it is validated that the developed scheme significantly outperforms the CL-based scheme.


Dual-Zone Hard-Core Model for RTS/CTS Handshake Analysis in WLANs

January 2025

IEEE Transactions on Wireless Communications

This paper introduces a new stochastic geometry-based model to analyze the Request-to-Send/Clear-to-Send (RTS/CTS) handshake mechanism in wireless local area networks (WLANs). We develop an advanced hard-core point process model, termed the dual-zone hard-core process (DZHCP), which extends traditional hard-core models to capture the spatial interactions and exclusion effects introduced by the RTS/CTS mechanism. This model integrates key parameters accounting for the thinning effects imposed by RTS/CTS, enabling a refined characterization of active transmitters in the network. Analytical expressions are derived for the intensity of the DZHCP, the mean interference, and an approximation of the success probability, providing insight into how network performance depends on critical design parameters. Our results demonstrate that the Type II RTS/CTS mechanism significantly reduces mean interference by introducing additional protection regions, effectively mitigating the impact of nearby interferers. Compared to CSMA-based schemes, it achieves a lower mean interference level while maintaining a comparable or higher active node density.


A Rate-Distortion Analysis for Composite Sources Under Subsource-Dependent Fidelity Criteria

January 2025

·

1 Read

·

1 Citation

IEEE Journal on Selected Areas in Communications

A composite source, consisting of multiple subsources and a memoryless switch, outputs one symbol at a time from the subsource selected by the switch. If some data should be encoded more accurately than other data from an information source, the composite source model is suitable because in this model different distortion constraints can be put on the subsources. In this context, we propose subsource-dependent fidelity criteria for composite sources and use them to formulate a rate-distortion problem. We solve the problem and obtain a single-letter expression for the rate-distortion function. Further rate-distortion analysis characterizes the performance of classify-then-compress (CTC) coding, which is frequently used in practice when subsource-dependent fidelity criteria are considered. Our analysis shows that CTC coding generally has performance loss relative to optimal coding, even if the classification is perfect. We also identify the cause of the performance loss, that is, class labels have to be reproduced in CTC coding. Last but not least, we show that the performance loss is negligible for asymptotically small distortion if CTC coding is appropriately designed and some mild conditions are satisfied.


Dual-Zone Hard-Core Model for RTS/CTS Handshake Analysis in WLANs

December 2024

·

3 Reads

This paper introduces a new stochastic geometry-based model to analyze the Request-to-Send/Clear-to-Send (RTS/CTS) handshake mechanism in wireless local area networks (WLANs). We develop an advanced hard-core point process model, termed the dual-zone hard-core process (DZHCP), which extends traditional hard-core models to capture the spatial interactions and exclusion effects introduced by the RTS/CTS mechanism. This model integrates key parameters accounting for the thinning effects imposed by RTS/CTS, enabling a refined characterization of active transmitters in the network. Analytical expressions are derived for the intensity of the DZHCP, the mean interference, and an approximation of the success probability, providing insight into how network performance depends on critical design parameters. Our results provide better estimates of interference levels and success probability, which could inform strategies for better interference management and improved performance in future WLAN designs.



Citations (44)


... In [13], accelerated versions of the classical Arimoto-Blahut algorithm are provided, by parametrizing the recurrence formula of classical algorithm, and their work demonstrated that the convergence is at least as fast as the Arimoto-Blahut algorithm [1]. Moreover, authors in [3] introduced a modification to the classical Arimoto-Blahut algorithm and examined the dual problem of computing the channel capacity, known as the rate-distortion problem. They reported a convergence rate of at least O(1/t) and derived that their algorithm requires at least 2 log(n)/ε iterations. ...

Reference:

Revisit the Arimoto-Blahut algorithm: New Analysis with Approximation
A Constrained BA Algorithm for Rate-Distortion and Distortion-Rate Functions
  • Citing Article
  • March 2025

CSIAM Transactions on Applied Mathematics

... Most existing studies utilizing machine learning approaches to design communication systems lack theoretical justification for their proposed methods, and typically regard learned communication systems as black boxes [6], [7], [8]. Recently, a few works are aiming to conduct theoretical analyses for learningbased communication systems [9], [10], [11]. Specifically, these studies leverage statistical learning theory [12], [13] to derive generalization bounds based on the Independently and Identically Distributed (I.I.D.) channel assumption. ...

PAC Learnability for Reliable Communication Over Discrete Memoryless Channels
  • Citing Conference Paper
  • July 2024

... The asymmetrically and symmetrically clipped optical ASCO-OFDM technique is another hybrid approach that achieves a higher data rate than ACO-OFDM and FLIP-OFDM, while keeping a comparable BER performance to DCO-OFDM. More details about these O-OFDM schemes can be found in papers [8]- [12]. ...

Information Theoretic Limits of Improved ACO-OFDM Receivers in Optical Intensity Channels With Time Dispersion

IEEE Open Journal of the Communications Society

... Some attempts have been made to connect IB to optimal transport (OT) theory. The work in [8] reformulates the IB paradigm as an entropic OT problem. The InfoOT framework in [9] maximizes kernelized mutual information by learning an OT plan associated with the joint distribution of variables. ...

Information Bottleneck Revisited: Posterior Probability Perspective with Optimal Transport

... For instance, the authors of [20] propose a plug-in estimator for the RDF based on the empirical distribution of the observed data. Recently, a potent alternative to BAA was explored in [21], which proposes an alternative to calculate the RDF by drawing an equivalence with the optimal transport problem and then utilizing the Sinkhorn algorithm [22]. Unfortunately, such methodologies do not scale with dimension when continuous spaces are considered due to the exponential increase in required bins. ...

A Communication Optimal Transport Approach to the Computation of Rate Distortion Functions

... Ordered Reliability Bits Guessing Random Additive Noise Decoding (ORBGRAND) [25] is a soft decision GRAND decoder that approximates rank-ordered bit reliabilities using piecewise linear functions, allowing for the efficient generation of putative noise sequences through generative integer partition algorithms. ORBGRAND is therefore computationally efficient and has been shown to nearly achieve the capacity in AWGN channels [26]. GRAND-EDGE and ORBGRAND-EDGE (Erasure Decoding by Gaussian Elimination) were introduced in [27] to address potential jamming effects. ...

ORBGRAND Is Almost Capacity-Achieving
  • Citing Article
  • January 2022

IEEE Transactions on Information Theory

... The vector case of GNND aided by successive decoding has been studied in [7], thereby extending the renowned vertical Bell Laboratories layered space-time (V-BLAST) receiver [16], [17] for multiple-input-multiple-output (MIMO) channels, wherein the linear minimum mean squared error (LMMSE)-successive interference cancellation (SIC) procedure turns out to be a special case of GNND when the receive side information of fading is perfect. A case of GNND for non-ergodic fading channels using outage probability as performance measure has been studied in [18]. ...

Linear Shrinkage Receiver for Slow Fading Channels under Imperfect Channel State Information
  • Citing Conference Paper
  • November 2022

... Chang proposed the construction of ternary perfect sequences of length 3 m −1 2 [2] and a large class of such ternary sequences with length p m −1 p−1 was constructed by Ipatov [15] using shift register sequences over finite field F p m , where p m is an odd prime power. Recently, for the case p = 3 and m being odd, Liu, Zhang, and Yang [23] determined the cross-correlation spectrum of Ipatov sequences and their 2-decimation sequences. Based on the theory of quadratic forms and exponential sums over finite fields, it was shown in [23] that the cross-correlation function takes on three low values. ...

Ternary perfect sequences with three-valued cross-correlation
  • Citing Article
  • January 2022

Advances in Mathematics of Communications

... Optimal transport theory [22,27,28] has been successfully applied in different fields [8,9,17,21,30,38,40,41,50,52]. Consequently, there are numerous algorithms for solving classical optimal transport problem proposed in different perspectives, such as linear programming methods [51], primal-dual algorithms [23], solving Monge-Ampère equation [5,13,26,39], proximal block coordinate descent methods [24], reduction and approximation techiniques for high-dimensional distributions [31,35,36,53], Sinkhorn algorithm and its variants [2,3,12,[32][33][34]42], etc. ...

An Optimal Transport Approach to the Computation of the LM Rate

... With the rapid growth of multimedia data traffic in modern intelligent information services, efficient data transmission has become a critical requirement for wireless communication systems. Semantic communication (SemCom) [2] has emerged as a transformative technology, offering significant potential to enhance communication efficiency and is widely recognized as a cornerstone of next-generation networks. Unlike traditional approaches, SemCom focuses on transmitting the intended meaning or core information of the data, thereby minimizing irrelevant and redundant information during transmission. ...

An Indirect Rate-Distortion Characterization for Semantic Sources: General Model and the Case of Gaussian Observation
  • Citing Article
  • September 2022

IEEE Transactions on Communications