ArticlePDF Available

Abstract

The deterministic identification (DI) capacity is developed in multiple settings of channels with power constraints. A full characterization is established for the DI capacity of the discrete memoryless channel (DMC) with and without input constraints. Originally, Ahlswede and Dueck established the identification capacity with local randomness at the encoder, resulting in a double exponential number of messages in the block length n . In the deterministic setup, the number of messages scales exponentially, as in Shannon’s transmission paradigm, but the achievable identification rates are higher. An explicit proof was not provided for the deterministic setting. In this paper, a detailed proof is presented for the DMC. Furthermore, Gaussian channels with fast and slow fading are considered, when channel side information is available at the decoder. A new phenomenon is observed as we establish that the number of messages scales as 2nlog(n)R2^{n\log (n)R} by deriving lower and upper bounds on the DI capacity on this scale. Consequently, the DI capacity of the Gaussian channel is infinite in the exponential scale and zero in the double exponential scale, regardless of the channel noise.
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 68, NO. 1, JANUARY 2022 1
Deterministic Identification Over Channels
With Power Constraints
Mohammad J. Salariseddigh ,UziPereg ,Member, IEEE, Holger Boche ,Fellow, IEEE,
and Christian Deppe ,Member, IEEE
Abstract The deterministic identification (DI) capacity is
developed in multiple settings of channels with power constraints.
A full characterization is established for the DI capacity of the
discrete memoryless channel (DMC) with and without input
constraints. Originally, Ahlswede and Dueck established the
identification capacity with local randomness at the encoder,
resulting in a double exponential number of messages in the block
length n. In the deterministic setup, the number of messages
scales exponentially, as in Shannon’s transmission paradigm,
but the achievable identification rates are higher. An explicit
proof was not provided for the deterministic setting. In this
paper, a detailed proof is presented for the DMC. Furthermore,
Gaussian channels with fast and slow fading are considered,
when channel side information is available at the decoder.
A new phenomenon is observed as we establish that the number
of messages scales as 2nlog(n)Rby deriving lower and upper
bounds on the DI capacity on this scale. Consequently, the DI
capacity of the Gaussian channel is infinite in the exponential
scale and zero in the double exponential scale, regardless of the
channel noise.
Index Terms—Channel capacity, identification, deterministic
codes, Gaussian fading channels, super exponential growth,
channel side information.
I. INTRODUCTION
IN THE fundamental communication paradigm considered
by Shannon [3], a sender wishes to convey a message
Manuscript received November 5, 2020; revised September 22, 2021;
accepted October 17, 2021. Date of publication October 26, 2021; date of
current version December 23, 2021. The work of Mohammad J. Salariseddigh,
Uzi Pereg, and Christian Deppe was supported by the Lehr- und Forschung-
seinheit für Nachrichtentechnik (LNT), New Communication Models-Post
Shannon Communication (NEWCOM) under Grant 16KIS1005. The work
of Holger Boche was supported in part by LTI, NEWCOM under Grant
16KIS1003K and in part by the Bundesministerium für Bildung und
Forschung (BMBF) within the National Initiative for Macroscopic Mole-
cular Communications (MAMOKO) under Grant 16KIS0914. An earlier
version of this paper was presented in part at the IEEE Information
Theory Workshop (ITW 2020) [DOI: 10.1109/ITW46852.2021.9457587],
in part at the IEEE International Conference on Communications (ICC
2021) [DOI: 10.1109/ICC42927.2021.9500406], in part at the Post-Shannon
Communication Session, London Symposium of Information Theory (LSIT
2021) (as a talk), and in part at the VDE-TUM Workshop on Molecu-
lar Communication and 6G in September 2021. (Corresponding author:
Mohammad J. Salariseddigh.)
Mohammad J. Salariseddigh, Uzi Pereg, and Christian Deppe are with the
Institute for Communications Engineering, Technische Universität München,
80333 Munich, Germany (e-mail: mjss@tum.de; uzi.pereg@tum.de;
christian.deppe@tum.de).
Holger Boche is with the Institute of Theoretical Information Technology,
Technische Universität München, 80290 Munich, Germany, also with the
Munich Center for Quantum Science and Technology (MCQST), 80799
Munich, Germany, and also with the Cyber Security in the Age of Large-
Scale Adversaries—Exzellenzcluster (CASA), 44780 Bochum, Germany
(e-mail: boche@tum.de).
Communicated by F. Alajaji, Associate Editor for Shannon Theory.
Color versions of one or more figures in this article are available at
https://doi.org/10.1109/TIT.2021.3122811.
Digital Object Identifier 10.1109/TIT.2021.3122811
through a noisy channel in such a manner that the receiver
will be able to retrieve the original message. In other words,
the decoder’s task is to determine which message was sent.
Ahlswede and Dueck [4] introduced a scenario of a different
nature, where the decoder only performs identification and
determines whether a particular message was sent or not
[4]–[6]. Applications include the tactile internet [7], vehicle-
to-X communications [8], [9], digital watermarking [10]–[12],
online sales [13], [14], industry 4.0 [15], health care [16], and
other event-triggered systems.
We give two motivating examples for applications of
the identification paradigm. Molecular communication is a
promising contender for future applications such as the sixth
generation of cellular communication (6G) [17], [18], in which
a number of applications require alerts to be identified [18].
Furthermore, in other systems of molecular communication,
a nano-device needs to determine the occurrence of a specific
event. For example, in the course of targeted drug deliv-
ery [19], [20] or cancer treatment [21]–[23], a nano-device
will seek to know whether the blood pH exceeds a critical
threshold or not, whether a specific drug is released or not,
whether another nano-device has replicated itself, whether a
certain molecule was detected, whether a target location in
the vessels is identified, or whether the molecular storage is
empty, etc., [24]. A second application for identification is
vehicle-to-X communications, where a vehicle that collects
sensor data may ask whether a certain alert message concern-
ing the future movement of an adjacent vehicle was transmitted
or not [25, Sec. VII].
The identification problem [4] can be regarded as a
Post-Shannon [26] model where the decoder does not perform
an estimation, but rather a binary hypothesis test to decide
between the hypotheses ‘sent or ‘not sent’, based on the
observation of the channel output. As the sender has no knowl-
edge of the desired message that the receiver is interested in,
the identification problem can be regarded as a test of many
hypotheses occurring simultaneously. The scenario where the
receiver misses and does not identify his message is called
a type I error, or ‘missed identification’, whereas the event
where the receiver accepts a false message is called a type II
error, or ‘false identification’.
Ahlswede and Dueck [4] required randomized coding for
their identification-coding scheme. This means that a random-
ized source is available to the sender. The sender can make his
encoding dependent on the output of this source. It is known
that this resource cannot be used to increase the transmission
capacity of discrete memoryless channels [27]. A remarkable
result of identification theory is that given local randomness
0018-9448 © 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See https://www.ieee.org/publications/rights/index.html for more information.
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
2IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 68, NO. 1, JANUARY 2022
at the encoder, reliable identification can be attained such
that the code size, i.e., the number of messages, grows
double exponentially in the block length n, i.e., 22nR [4].
This differs sharply from the traditional transmission setting
where the code size scales only exponentially, i.e., 2nR.
Beyond the exponential gain in identification, the extension
of the problem to more complex scenarios reveals that the
identification capacity has a very different behavior compared
to the transmission capacity [28]–[33]. For instance, feedback
can increase the identification capacity [28] of a memory-
less channel, as opposed to the transmission capacity [34].
Nevertheless, it is difficult to implement randomized-encoder
identification (RI) codes that will achieve such performance,
because it requires the encoder to process a bit string of
exponential length. The construction of identification codes
is considered in [6], [35]–[38]. Identification for Gaussian
channels is considered in [30], [39]–[42].
In the deterministic setup for a DMC, the number of mes-
sages scales exponentially in the blocklength [4], [43]–[45],
as in the traditional setting of transmission. Nevertheless,
the achievable identification rates are significantly higher
than those of transmission. In addition, deterministic codes
often have the advantage of simpler implementation and
simulation [46], explicit construction [47], and single-block
reliable performance. In particular, JáJá [44] showed that
the deterministic identification (DI) capacity1of a binary
symmetric channel is 1 bit per channel use, based on com-
binatorial methods. The meaning of this result is that one
can exhaust the entire input space and assign (almost) all
sequences in the n-dimensional space {0,1}nas codewords.
Ahlswede et al. [4], [43] stated that the DI capacity of a
discrete memoryless channel (DMC) with a stochastic matrix
Wis given by the logarithm of the number of distinct row
vectors of W(see Section IV. in [4] and abstract of [43]).
Nonetheless, an explicit proof for this result was not provided
in [4], [43]. Instead, Ahlswede and Cai [43] referred the reader
to a paper [48] which does not include identification and
addresses a completely different model of an arbitrarily vary-
ing channel [48]. Since then, the problem of proving this result
has remained unsolved, since a straightforward extension of
the methods in [48], using decoding territories, does not seem
to yield the desired result on the DI capacity [49].
Modern communications require the transfer of enormous
amounts of data in wireless systems, for cellular commu-
nication [37], sensor networks [50], smart appliances [51],
and the internet of things [52], etc. Wireless communication
is often modelled by fading channels with additive white
Gaussian noise [53]–[61]. In the fast fading regime, the
transmission spans over a large number of coherence time
intervals [62], hence the signal attenuation is characterized
by a stochastic process or a sequence of random parameters
[63]–[66]. In some applications, the receiver may acquire
channel side information (CSI) by instantaneous estimation of
the channel parameters [67]–[69]. On the other hand, in the
slow fading regime, the latency is short compared to the
1The DI capacity in the literature is also referred to as the non-randomized
identification (NRI) capacity [43] or the dID capacity [30].
coherence time [62], and the behaviour is that of a compound
channel [70]–[74].
In this paper, we establish the DI capacity of channels
subject to an input constraint. Such a constraint is often
associated with a limited power supply or regulation, as in
the case of the Gaussian channel. We consider the settings of
a DMC and of Gaussian channels with fast fading and slow
fading, with CSI available at the decoder. For a DMC, one may
assume without loss of generality that the rows of the channel
matrix are distinct (see Remarks 3 and 5). Our first result
is that the DI capacity of a DMC Wunder this assumption,
subject to the input constraint 1
nn
t=1 φ(xt)A,isgivenby
CDI(W)= max
pX:E{φ(X)}≤AH(X).(1)
We note that the DI capacity does not depend on the
specific values of the transition probabilities, as long as the
rows of the channel matrix are distinct. This result has the
following geometric interpretation. At rst glance, it may
seem reasonable that for the purpose of identification, one
codeword could represent two messages. While identification
allows overlap between decoding regions [75], [76], it turns
out that overlap at the encoder is not allowed for deterministic
codes. However, if two messages are represented by the same
codeword, then the low probability of a type I error comes
at the expense of the high probability of a type II error, and
vice versa. Thus, DI coding imposes the restriction that every
message must have a distinct codeword. The converse proof
follows from this property in a straightforward manner, since
the volume of the input subset of sequences that satisfy the
input constraint is 2nCDI(W). A similar principle guides the
direct part as well. The input space is covered such that each
codeword is surrounded by a sphere of radius to separate
the codewords.
Gaussian channels have an entirely different and unusual
behaviour. We consider deterministic identification for
Gaussian channels with fast fading and slow fading, and
with channel side information (CSI) available at the decoder.
Applying discretization to our capacity result above (see (1)),
we obtain that the DI capacity of the standard Gaussian
channel is infinite in the exponential scale (as we have recently
observed in [2], [77]). However, for a finite blocklength n,the
number of codewords must be finite. Thereby, the meaning
of the infinite capacity result is that the number of messages
scales super-exponentially. This raises the question: What is
thetrueorderofthecodesize.In mathematical terms, what
is the scale L(n, R)for which the DI capacity is positive yet
finite. We show that for Gaussian channels, the number of
messages scales as 2nlog(n)R, and develop lower and upper
bounds on the DI capacity in this scale. As a consequence,
we deduce that the DI capacity of a Gaussian channel with
fast fading is infinite in the exponential scale and zero in the
double exponential scale, regardless of the channel noise. For
slow fading, the DI capacity in the exponential scale is infinite,
unless the fading gain can be zero or arbitrarily close to zero
(with positive probability), in which case the DI capacity is
zero. Note, however, that this scale is comparably lower than
the double exponential scale of RI coding.
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
SALARISEDDIGH et al.: DETERMINISTIC IDENTIFICATION OVER CHANNELS WITH POWER CONSTRAINTS 3
By providing a detailed proof for the DMC, we thus fill the
gap in the previous analysis [4], [43] as well. In the proof,
we use the method of types, while the derivation is based
on ideas that are analogous to the combinatoric analysis of
Hamming distances by JáJá [44]. Although the codebook con-
struction is similar to that of Ahlswede’s coding scheme [48],
the decoder is significantly different. In particular, we do not
use decoding territories as in [48], but rather perform a typical-
ity check. Nonetheless, the type-class intersection lemma and
the message-set analysis in [48] turn out to be useful in our
analysis as well. Hence, our proof combines techniques and
ideas from both works, by JáJá [44] and by Ahlswede [48],
to derive the DI capacity both with and without an input
constraint. The analysis for Gaussian channels also relies
on geometric considerations, using sphere packing. Based
on fundamental properties of packing arrangements [78], the
optimal packing of non-overlapping spheres of radius
contains an exponential number of spheres, and by decreasing
the radius of the codeword spheres, the exponential rate can be
made arbitrarily large. However, in the derivation of our lower
bound in the 2nlog(n)R-scale, we pack spheres of a sub-linear
radius nn1/4, which results in 21
4nlog(n)codewords.
The paper is organized as follows. In Section II we give the
definitions and a brief review of related work. In Section III
we address deterministic identification for the DMC with and
without an input constraint. In Subsection III-A, a channel
reduction procedure is described such that high-cost identical
rows are removed from the channel matrix. The capacity
theorem is stated in Subsection III-B. The direct part is
proved in Subsection III-C, and the converse proof is in
Subsection III-D. We consider deterministic identification over
fading channels in Section IV. Definitions for fading channels
are given in Subsection IV-A and IV-B.
Lower and upper bounds in the 2nlog(n)R-scale are pre-
sented in Subsection IV-C, and the respective proofs in IV-D
and IV-E. Section V is dedicated to summary and discussion.
II. DEFINITIONS AND RELATED WORKS
In this section we introduce the channel model and coding
definitions. Here we only consider the discrete memoryless
channel (DMC). The channel description and coding definition
for the Gaussian channel will be presented in Section IV.
A. Notation
We use the following notation conventions throughout. Cal-
ligraphic letters X,Y,Z,... are used for finite sets. Lowercase
letters x,y,z,... stand for constants and values of random
variables, and uppercase letters X,Y,Z,... stand for random
variables. The distribution of a random variable Xis specified
by a probability mass function (pmf) pX(x)over a finite
set X. The set of all pmfs over Xis denoted by P(X).
H(X)and I(X;Y)are the entropy and mutual information,
respectively; H2(p)=(1 p) log(1 p)plog(p)is the
binary entropy function for a given 0<p<1; all logarithms
and information quantities are taken to the base 2.Weuse
xj=(x1,x
2,...,x
j)to denote a sequence of letters from
X. A random sequence Xnand its distribution pXn(xn)are
defined accordingly. The set of consecutive natural numbers
from 1through Mis denoted by [[ M]] . The Hamming distance
between two sequences anand bnis defined as the number
of positions for which the sequences have different symbols,
i.e., dH(an,b
n)=|{t[[ n]] ; at=bt}|. The notation x=
(x1,x
2,...,x
n)is used instead of xnwhen it is understood
from the context that the length of the sequence is n,andthe
2-norm of xis denoted by x.Then-dimensional Hamming
sphere of radius that is centered at anis defined as
Sε(an)={xn∈Xn:dH(xn,a
n)<nε}.(2)
We denote the hyper-sphere of radius raround x0by
Sx0(n, r)=xRn:xx0≤r,(3)
and its volume by Vol(S). The closure of a set Ais denoted
by cl(A). In the continuous case, we use the cumulative
distribution function FX(x)=Pr(Xx)for xR,
or alternatively, the probability density function (pdf) fX(x),
when it exists. The element-wise product of vectors is denoted
by xy=(xtyt)n
t=1. A random sequence Xand its
distribution FX(x)are defined accordingly.
B. Channel Description
ADMC(X,Y,W)consists of finite input and output
alphabets Xand Y, respectively, and a conditional pmf
W(y|x). The channel is memoryless without feedback, and
therefore Wn(yn|xn)=n
t=1 W(yt|xt). We denote a DMC
by W=(X,Y,W). Next, we consider an input constraint.
Let φ:X→[0,)be some given bounded cost function,
and define
φn(xn)=1
n
n
t=1
φ(xt).(4)
Given an input constraint A>0corresponding to the cost
function φn(xn), the channel input xnmust satisfy
φn(xn)A. (5)
We may assume without loss of generality that 0A
φmax,whereφmax =max
x∈X φ(x).Itisalsoassumedthatfor
some x0∈X,φ(x0)=0.
C. Coding
The definitions for DI codes, achievable rates, and capacity
are given below. In this paper we consider codes with different
size orders. For instance when we discuss the exponential
scale, we refer to a code size that scales as L(n, R)=2
nR.
On the other hand, in the double exponential scale, the code
size is L(n, R)=2
2nR . Later, in Section IV where we con-
sider Gaussian channels, we will see that the appropriate scale
turns out to be neither exponential nor double exponential, but
in between. Throughout the paper we use the mathematical
convention that L:N×R+Ndenotes a map, and
L(n, R)Nis its value for a given blocklength nand
rate R.
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
4IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 68, NO. 1, JANUARY 2022
Definition 1: Let L1(n, R)and L2(n, R)be two coding
scales. We say that L1dominates L2if
lim
n→∞
L2(n, b)
L1(n, a)=0,(6)
for all a, b > 0. We will denote this relation by L2L1.
In complexity theory of computer science, the relation
above is denoted by the ‘small o-notation’, L2(n, 1) =
o(L1(n, 1)) [79]. Beyond exponential, other orders that com-
monly appear in complexity theory are the linear, logarithmic,
and polynomial scales, nR,log(nR),and(nR)k. The corre-
sponding ordering is
log(nR)nR (nR)k2nR 2nlog(n)R22nR .(7)
Definition 2: An (L(n, R),n)DI code for a DMC W
under input constraint A, assuming L(n, R)is an integer,
is defined as a system (U,D)that consists of a codebook
U={ui}i[[ L(n,R)]],U⊂X
n, such that
φn(ui)A, for all i[[ L(n, R)]],(8)
and a collection of decoding regions D={Di}i[[ L(n,R)]]
with L(n,R)
i=1 Di⊂Y
n. Given a message i[[ L(n, R)]],
the encoder transmits ui. The decoder’s aim is to answer
the following question: Was a desired message jsent or
not? Two types of errors may occur: Rejecting of the true
message, or accepting a false message. Those error events are
often referred to as type I and type II errors, respectively.
Specifically, P(n)
e,1(i)is the type I error probability for rejecting
thetruemessagei, while P(n)
e,2(i, j)is the type II error
probability for accepting the false message j, given that the
message iwas sent.
The error probabilities of the identification code (U,D)are
given by
Pe,1(i)=Wn(Dc
i|ui)(missed-identification error),(9)
Pe,2(i, j)=Wn(Dj|ui)(false identification error).(10)
An (L(n, R),n
1
2)DI code further satisfies
Pe,1(i)λ1,(11)
Pe,2(i, j)λ2,(12)
for all i, j [[ L(n, R)]] such that i=j.
ArateR>0is called achievable if for every λ1
2>0
and sufficiently large nthere exists an (L(n, R),n
1
2)DI
code. The operational DI capacity is defined as the supremum
of achievable rates and will be denoted by CDI (W,L).
As mentioned earlier, Ahlswede and Dueck [4] needed
randomized encoding for their identification-coding scheme.
This means that a randomized source is available to the
sender. The sender can make his encoding dependent on
the output of this source. Therefore, a randomized-encoder
identification (RI) code is defined in a similar manner where
the encoder is allowed to select a codeword Uiat random,
according to some conditional input distribution Q(xn|i).The
RI capacity is then denoted by CRI (W,L). Given local ran-
domness at the encoder, reliable identification can be attained
such that the number of messages grows double exponentially
Fig. 1. Geometric illustration of identification errors in the deterministic
setting. The arrows indicate three scenarios for the channel output, given that
the encoder transmitted the codeword u1corresponding to i=1.Ifthe
channel output is outside D1, then a type I error has occurred, as indicated
by the bottom red arrow. This kind of error is also considered in traditional
transmission. In identification, the decoding sets can overlap. If the channel
output belongs to D1but also belongs to D2, then a type II error has occurred,
as indicated by the middle brown arrow. Correct identification occurs when
the channel output belongs only in D1, which is marked in blue.
in the block length n,i.e.,L(n, R)=Ldouble(n, R)
22nR [4]. This differs sharply from the traditional transmis-
sion setting where the code size scales only exponentially,
i.e., L(n, R)=Lexp(n, R)2nR . Remarkably, in [4] it
was shown that CRI (W,L
double)=CT(W,L
exp),where
CT(W,L
exp)denotes the transmission capacity of the channel
in the exponential scale.
Remark 1: The code scale can also be thought of as a
sequence of monotonically increasing functions Ln(R)of the
rate. Hence, given a code of size M=Ln(R), the coding
rate can be obtained from the inverse relation R=L1
n(M).
In particular, for the transmission setting [3], or DI coding for
a DMC [44], the coding rate is defined as
R=1
nlog(M).(13)
Whereas for RI coding [4], the rate was defined as
R=1
nlog log(M).(14)
On the other hand, using the scale L(n, R)=2
nlog(n)Ras
for Gaussian channels stated in Theorem 10, the coding rate
is
R=log M
nlog n.(15)
Remark 2: It can be readily shown that in general, if the
capacity in an exponential scale is finite, then it is zero in
the double exponential scale. Conversely, if the capacity in a
double exponential scale is positive, then the capacity in the
exponential scale is +. This principle can be generalized to
any pair of scales L1and L2,whereL2is dominated by L1.
We come back to this in Subsection IV-C.
A geometric illustration for the type I and II error prob-
abilities is given in Figure 1. When the encoder sends the
message ibut the channel output is outside Di, then type I
error occurs. This kind of error is also considered in traditional
transmission. In identification, the decoding sets can overlap.
A type II error covers the case where the output sequence
belongs to the intersection of Diand Djfor j=i.
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
SALARISEDDIGH et al.: DETERMINISTIC IDENTIFICATION OVER CHANNELS WITH POWER CONSTRAINTS 5
D. Related Work
We briefly review Ahlswede and Dueck’s result [4] on
the RI capacity, i.e., when the encoder uses a stochastic
mapping. As mentioned above, using RI codes, it is possible
to identify a double exponential number of messages in the
block length n. That is, given a rate R<CRI (W,L),there
exists a sequence of (L(n, R)=2
2nR ,n)RI codes with
vanishing error probabilities. Despite the significant difference
between the definitions in the identification setting and in the
transmission setting, it was shown that the value of the RI
capacity in the double exponential scale equals the Shannon
capacity of transmission.
Theorem 1 (See [4], [80]): The RI capacity in the double
exponential scale of a DMC Wis given by
CRI (W,L)= max
pX:E{φ(X)}≤AI(X;Y),for L(n, R)=2
2nR .
(16)
Hence, the RI capacity in the exponential scale is infinite,
i.e.,
CRI (W,L)=,for L(n, R)=2
nR.(17)
In the next sections, we will consider the identification set-
ting when the encoder does not have access to randomization.
Theorem 2 (See [4], [40]): Let Gdenote the standard
Gaussian channel, where the input-ouptut relation is given by
Y=gX +Z, with Z∼N(0
2
Z)and a fixed known gain
g>0. Then, the RI capacity in the double exponential scale
is given by
CRI (G,L)=1
2log 1+g2A
σ2
Z,for L(n, R)=2
2nR .
(18)
Hence, the RI capacity in the exponential scale is infinite,
i.e.,
CRI (G,L)=,for L(n, R)=2
nR.(19)
III. MAIN RESULTS—DMC
We give our main results on the DI capacity of the DMC.
For a DI code, as opposed to the randomized case, the number
of messages 2nR is only exponential in the blocklength. In this
sense, DI codes are similar to transmission codes. However,
the achievable rates for identification are significantly higher,
as the DI capacity is given in terms of the input entropy instead
of the mutual information.
A. Channel Reduction
We begin with a procedure of channel reduction where we
remove identical rows from the channel matrix, so that the
remaining input letters have a lower cost compared to the
deleted letters. As will be seen below, the DI capacity remains
the same following this reduction. The characterization of the
DI capacity will be given in the next section in terms of the
reduced input alphabet.
We begin with the definition of the reduced channel.
Definition 3 (Reduced Channel): Given a DMC Wwith
a stochastic matrix W:X→Y, we define the reduced
DMC Wras follows. Let {X()}be a partition of Xinto
equivalent classes, so that two letters xand xbelong to the
same equivalent class if and only if the corresponding rows
are identical, namely
x, x∈X()W(y|x)=W(y|x)y∈Y.(20)
For every class X(), assign a representative element
z()=arg min
x∈X()φ(x),(21)
which is associated with the lowest input cost. If there is more
than one letter that is associated with the lowest input cost in
X(), then choose one of them arbitrarily. Then the reduced
input alphabet is defined as
Xr={z()},(22)
and the reduced DMC Wris defined by a channel matrix
Wr:Xr→Y, consisting of the rows in Xr,i.e.,
Wr(y|x)=W(y|x),(23)
for x∈X
rand y∈Y.
Lemma 3: The operational capacities of the reduced channel
Wrand the original channel Ware the same:
CDI (W,L)=CDI(Wr,L),for L(n, R)=2
nR.(24)
We give the proof of Lemma 3 in Appendix A. As we will
see shortly, the DI capacity of a DMC Wdepends on Wonly
through Xr. That is, the DI capacity does not depend on the
individual values of the channel matrix and depends solely on
the distinctness of its rows.
Remark 3: Based on Lemma 3, it is sufficient to consider
a channel with distinct rows. That is, if we establish the
DI capacity for channels with distinct rows, we can then
determine the DI capacity for a general channel. In other
words, in order to derive a capacity result, we may assume
without loss of generality that the channel rows are distinct.
B. Capacity Theorem
In this section, we give our main result on the DI capacity
of a channel subject to input constraint. The capacity result
is stated in terms of the reduced channel as defined in the
previous section. Let Wbe a DMC channel with input cost
function φ(x)and input constraint Aas specified in (5). Define
CDI (W)= max
pX:E{φ(X)}≤AH(X),(25)
for XpX.
Theorem 4: The DI capacity of a DMC Wunder input
constraint is given by
CDI(W,L)=CDI(Wr),for L(n, R)=2
nR,(26)
where Wrdenotes the reduced channel (see Definition 3).
Hence, the DI capacity in the double exponential scale is zero.
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
6IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 68, NO. 1, JANUARY 2022
We prove the direct part in Subsection III-C and the
converse part in Subsection III-D. As can be seen in Sub-
section III-D, the strong converse property holds for the DI
capacity [5, see Def. 3.5.1]. Notice that we have characterized
the DI capacity of the DMC Win terms of its reduced version,
as specified in Lemma 3.
Corollary 5 (Also in [4], [43]): The DI capacity of a DMC
Wwithout constraints, i.e., with A=φmax,isgivenby
CDI (W,L)=lognrow(W),(27)
for L(n, R)=2
nR where nrow(W)is the number of distinct
rows of W.
The corollary above is an immediate consequence of The-
orem 4. Indeed, for A=φmax,wehave
CDI (Wr)= max
pX0∈P(Xr)H(X0)
=log|Xr|
=lognrow(W),(28)
where X0is a random variable, the support of which is in
the reduced input alphabet Xras defined in Definition 3. The
second equality holds since the maximal value of H(X)is
log |X|, and the last equality because the size of the reduced
input alphabet is |Xr|=nrow(W).
Remark 4: Ahlswede et al. [4], [43] stated the result in
Corollary 5 on the DI capacity of a DMC without constraints
(see Section IV. [4] and abstract of [43]), without providing
an explicit proof. A straightforward extension of the methods
in [48], using decoding territories, does not seem to yield the
desired result on the DI capacity. Thereafter, the proof has
remained an open problem.
Remark 5: An alternative expression for the DI capacity is
as follows,
CDI(W,L)= max
pX0∈P(Xr): E{φ(X0)}≤AH(X0),(29)
where X0is as in (28). As explained in Remark 3, one may
assume without loss of generality that the channel has distinct
rows. Under this assumption, the DI capacity formula reduces
to the formula in (1), i.e.,
CDI(W,L)= max
pX:E{φ(X)}≤AH(X),(30)
for a channel Wwith distinct rows and L(n, R)=2
nR.
To illustrate our results, we give the following example.
Example 1: Consider the binary symmetric channel (BSC),
Y=X+Zmod 2,(31)
where X=Y={0,1},ZBernoulli(ε), with crossover
probability 0ε1
2. Suppose that the channel is subject to
a Hamming weight input constraint,
1
n
n
t=1
xtA, (32)
with φ(x)=x. Observe that for ε=1
2, the rows of the channel
matrix are identical. Hence, the reduced input alphabet consists
of one letter, and the DI capacity is zero (see Definition 3).
Fig. 2. The deterministic identification (DI) capacity of the BSC as a function
of the input constraint A. The dashed red line indicates the binary entropy
function, which is maximized in (33). The solid blue line indicates the DI
capacity.
Now, suppose that ε< 1
2. Then the rows of the channel
matrix W=1εε
ε1εare distinct, hence Wr=W.
By Theorem 4, the DI capacity is given by
CDI(W,L)=CDI(W)= max
0pAH2(p),(33)
since the channel input is binary, where H2(p)=(1
p) log(1 p)plog(p)is the binary entropy function. There-
fore, the DI capacity of the BSC with Hamming weight
constraint is
CDI (W,L)=H2(A)if A<1
2
1if A1
2
,for L(n, R)=2
nR.
(34)
(See Figure 2). To show the direct part, set
XBernoulli(A)if A< 1
2and XBernoulli1
2,
otherwise. The converse part follows as the binary entropy
function H2(p)is strictly increasing on 0p1
2, attaining
its maximum value H2(1
2)=1, and strictly decreasing on
1
2<p1(see Figure 2). The geometric interpretation is
that the binary Hamming ball of radius np can be covered
with codewords. As the volume of the Hamming ball
is approximately 2nH2(p), one can achieve rates that are
arbitrarily close to H2(p). Without an input constraint, i.e.,
for A=1, we recover the result of JáJá [44],
CDI(W,L)=1.(35)
This example demonstrates that the DI capacity is discon-
tinuous in the channel statistics, as CDI (W,L)=1for ε< 1
2
and CDI(W,L)=0for ε=1
2.
C. Achievability Proof
Consider a DMC W. By Lemma 3 we can assume without
loss of generality that the channel matrix W:X→Yhas
distinct row vectors. To prove achievability of the DI capacity,
we combine methods and ideas from the work of JáJá [44] as
well as techniques by Ahlswede [48]. The analysis for the type
II error is based on ideas that are analogous to the combinatoric
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
SALARISEDDIGH et al.: DETERMINISTIC IDENTIFICATION OVER CHANNELS WITH POWER CONSTRAINTS 7
analysis of Hamming distances in [44]. The codebook con-
struction is similar to that of Ahlswede’s coding scheme [48],
yet the decoder is significantly different. Nonetheless, the type-
class intersection lemma and the message-set analysis in [48]
are useful in our analysis for the type II error.
We extensively use the method of types [70, Ch. 2]. Here
a brief review of the definitions for type classes and δ-typical
sets is given. The type ˆ
Pxnof a given sequence xnis defined
as the empirical distribution ˆ
Pxn(a)=N(a|xn)/n for a∈X,
where N(a|xn)is the number of occurrences of the symbol
a∈Xin the sequence xn. The space of all types over X
of sequences of length nis denoted by Pn(X).Theδ-typical
set Tδ(pX)is defined as the set of sequences xn∈Xnsuch
that for every a∈X:|ˆ
Pxn(a)pX(a)|≤δif pX(a)>0,
and ˆ
Pxn(a)=0if pX(a)=0. A type class is denoted by
T(ˆ
P)={xn:ˆ
Pxn=ˆ
P}. Similarly, a joint type is denoted
by ˆ
Pxn,yn(a, b)=N(a, b|xn,y
n)/n for (a, b)∈X×Y,where
N(a, b|xn,y
n)is the number of occurrences of the symbol pair
(a, b)in the sequence (xi,y
i)n
i=1, and as a conditional type by
ˆ
Pyn|xn(b|a)=N(a, b|xn,y
n)/N (a|xn). The conditional δ-
typical set Tδ(pY|X|xn)is defined as the set of sequences yn
Ynsuch that for every b∈Y:|ˆ
Pyn|xn(b|a)pY|X(b|a)|≤δ
if pX,Y (a, b)>0,andpX,Y (a, b)=0if pX(a)=0.
1) The Codebook: First, we show that there exists a code
such that the codewords are separated by a distance of .Let
pX(x)be an input distribution on X, such that
E{φ(X)}=
x∈X
pX(x)φ(x)Aε(δ)(36)
for XpX(x),whereε(δ)0as δ0. We may assume
without loss of generality that pXis a type, due to the entropy
continuity lemma [70, Lem. 2.7].
Lemma 6: Let R<H(X). Then, for sufficiently small
ε(0,1) and sufficiently large n, there exists a codebook
U={vi,i∈M}, which consists of |M| sequences in Xn,
such that the following hold:
1) All the codewords belong to the type class T(pX),
namely
vi∈T(pX)for all i∈M.(37)
2) The codewords are distanced by , i.e.,
dH(vi,v
j) for all i=j. (38)
3) The codebook size is at least 1
2·2nR,thatis,
|M| 2n(R1
n).
Proof of Lemma 6: Denote
M2nR.(39)
Let U1,...,U
Mbe independent random sequences, each
uniformly distributed over the type class of pX,i.e.,
Pr (Ui=xn)=1
|T (pX)|xn∈T(pX),
0xn/∈T(pX).(40)
Next, define a new collection of sequences V1,...,V
Mas
follows,
Vi=Uiif dH(Ui,U
j) i=j,
otherwise,(41)
where dH(·,·)denotes the Hamming distance, and represents
an idle sequence of no interest. The assignment Vi=
is interpreted as “dropping the ith word Ui.” Consider the
following message set,
M={i:Vi=,i[[ M]] },(42)
corresponding to words that were not dropped, where we use
the notation
Mto indicate that the set is random.
We show that even though we removed words from the
original collection {Ui}i[[ M]] (of size M), the rate decrease
can be made negligible. Following the lines of [48], we derive
an upper-bound on Pr(|
M| 1
2M)where
Mdefined in (42)
is the operational message set. To this end, we will use the
following concentration lemma,
Lemma 7 (Also in [48]): Let A1,...,A
Kbe a sequence of
discrete random variables. Then,
Pr 1
K
K
i=1
Aic
2cK
K
i=1
max
ai1
E2AiAi1=ai1.(43)
Now, define an indicator for dropping the ith word by
ˆ
Vi=1Vi=,
0Vi=,(44)
and notice the equivalence between the following events,
M1
2M=M
i=1
ˆ
Vi>1
2M.(45)
Observe that ˆ
Vi=1, if and only if Uiis inside an ε-
sphere of some other Uj. Namely, ˆ
Vi=1iff Ui
j=iSε(Uj).
The selection of codewords can be viewed as an iterative
procedure. Specifically, define
Ai=
1Ui
j<iSε(Uj),
0otherwise,
(46)
Bi=
1Ui
j>iSε(Uj),
0otherwise.
(47)
Now, since ˆ
Vi=1implies that either Ai=1or Bi=1,
it follows that the number of dropped messages is bounded by
M
M=
M
i=1
ˆ
Vi
M
i=1
Ai+
M
i=1
Bi.(48)
Consider the event that
M
i=1
ˆ
Vi>1
2M. (49)
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
8IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 68, NO. 1, JANUARY 2022
If this holds, then the two sums in the right hand side of (48)
cannot be smaller than 1
4Mtogether, that is, either M
i=1 Ai
1
4M,orM
i=1 Bi1
4M, or both. Hence,
M
i=1
ˆ
Vi>1
2M
M
i=1
Ai1
4MM
i=1
Bi1
4M,(50)
and by the union bound,
Pr M
i=1
ˆ
Vi>1
2M
Pr M
i=1
Ai1
4M+PrM
i=1
Bi1
4M
=2Pr
M
i=1
Ai1
4M,(51)
where the last line follows by symmetry, as the random
variables ¯
A=M
i=1 Aiand ¯
B=M
i=1 Bihave the same
probability distribution.
Next we apply Lemma 7,
Pr M
i=1
Ai1
4M
21
4M
M
i=1
max
ai1
E2Ai|Ai1=ai1.(52)
Consider the conditional expectation above. Using the law
of total expectation, we can add conditioning on Ui1as well,
i.e.
E2Ai|Ai1=ai1
=
ui1
Pr(Ui1=ui1|Ai1=ai1)
·E(2Ai|Ui1=ui1,A
i1=ai1)
=
ui1
Pr(Ui1=ui1|Ai1=ai1)
·E(2Ai|Ui1=ui1)
max
ui1
E(2Ai|Ui1=ui1),(53)
where the second equality holds since Ai, is a deterministic
function of Ui1(see (46)). Hence, by (52)-(53),
Pr M
i=1
Ai1
4M21
4M
M
i=1
max
ui1
E2Ai|Ui1=ui1
=2
1
4M
M
i=1
max
ui11·Pr Ai=0|Ui1=ui1
+2·Pr Ai=1|Ui1=ui1
21
4M
M
i=1 1+2·max
ui1Pr(Ai=1|Ui1=ui1).
(54)
We bound the probability term Pr(Ai=1|Ui1=ui1),
as follows. For a Hamming sphere of radius ,
|Sε(xn)|≤n
·|X|
2(ε),(55)
for sufficiently large n,where
θ(ε)=H2(ε)+εlog |X|,(56)
tends to zero as ε0. The first inequality holds by a
simple combinatoric argument. Namely, counting the number
of sequences with up to different entries compared to a
given xn,wehaven
optional choices for the locations of
those entries, and |X| possible values for each of those entries.
The last inequality follows from Stirling’s approximation [81,
Example 11.1.3]. Hence,
M
j=1 Sε(uj)M2(ε)
=2
n(R+θ(ε)),(57)
for every given collection of sequences, u1,...,u
M∈T(pX).
Consider a random sequence ¯
Xnthat is uniformly distributed
over the type class T(pX), and statistically independent of
U1,...,U
M. We use this external sequence as an auxiliary in
the derivation below. Then,
Pr Ai=1Ui1=ui1
=Pr
Ui
j<i Sε(uj)
=Pr
¯
Xn
j<i Sε(uj)
Pr
¯
Xn
M
j=1 Sε(uj)
.(58)
The first equality follows from the definition of Aiin (46)
and because U1,...,U
Mare statistically independent. The
second equality holds because Uiand ¯
Xnare both uniformly
distributed over the type class of pX. The inequality follows
as Pr(F1)Pr(F1∪F
2)for every pair F1,F2of proba-
bilistic events. Since ¯
Xnis uniformly distributed over T(pX),
we have
Pr
¯
Xn
M
j=1 Sε(uj)
=
xn∈T (pX)
M
j=1 Sε(uj)
1
|T (pX)|
=1
|T (pX)|·T(pX)
M
j=1 Sε(uj)
2n(R+θ(ε))
|T (pX)|
(n+1)
|X| ·2n(R+θ(ε))
2nH(X)
2n(H(X)R2θ(ε)),(59)
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
SALARISEDDIGH et al.: DETERMINISTIC IDENTIFICATION OVER CHANNELS WITH POWER CONSTRAINTS 9
for sufficiently large n, where the first inequality follows
from (57), and the second is due to standard type class
properties [81, Th. 11.1.3]. The last expression tends to zero
as n→∞, provided that
R<H(X)3θ(ε).(60)
Together with (58)-(59), this implies
Pr Ai=1Ui1=ui12(ε).(61)
Now plugging (61) into (54) yields
Pr M
i=1
Ai1
4M21
4M1+2·2(ε)M
=21
4+23
4·2(ε)M,(62)
for sufficiently large n,wehave23
4·2(ε)25hence,
21
4+23
4·2(ε)21
4+2
5
=0.8721
<1.(63)
Thus we have a double exponential bound
Pr
M1
2M2α1M
=2
α12nR ,(64)
for some α1>0. We deduce that there exists at least one
codebook with the desired properties. This completes the proof
of Lemma 6.
We continue to the main part of the achievability proof. Let
U={vi,i∈M}be a codebook of size 2n(R1
n)as in
Lemma 6. Consider the following DI coding scheme for W.
a) Encoding: Given a message i∈Mat the sender,
transmit xn=vi.
b) Decoding: Let δ>0, such that δ0as ε0.Let
j∈Mbe the message that the decoder wishes to identify.
To do so, the decoder checks whether the channel output yn
belongs to the corresponding decoding set Djor not, where
Dj={yn:(vj,y
n)∈T
δ(pXW)}.(65)
Namely, given the channel output yn∈Y
n,if(vj,y
n)
Tδ(pXW), then the decoder declares that the message jwas
sent. On the other hand, if (vj,y
n)/∈T
δ(pXW), it declares
that jwas not sent.
2) Error Analysis: First, consider the error of type I, i.e.,
the event that Yn/∈D
i.Foreveryi∈M, the probability
of identification error of type I, Pe,1(i)=Pr((vi,Yn)/
Tδ(pXW)) tends to zero by standard type class considerations
[82, Th. 1.2].
We move to the error of type II, i.e., when Yn∈D
j
for j=i. To bound the probability of error Pe,2(i, j),
we use the conditional type-class intersection lemma, due to
Ahlswede [48], as stated below.
Lemma 8 (See [48, Lem. I1]): Let W:X→Ybe a
channel matrix of a DMC Wwith distinct rows. Then, for
every xn,x
n∈T
δ(pX)with dH(xn,x
n),
|Tδ(pY|X|xn)∩T
δ(pY|X|xn)|
|Tδ(pY|X|xn)|2nL(ε),(66)
with pY|XW, for sufficiently large nand some positive
function L(ε)>0which is independent of n.
Now, for short notation, denote the conditional δ-typical set
in Yn,givenxn∈T(pX),by
G(xn)≡T
δ(W|xn)={yn:(xn,y
n)∈T
δ(pXW)}.(67)
Then, for every i=j,
Pe,2(i, j)=Pr(Dj|xn=vi)
=
yn∈G(vj)
Wn(yn|vi)
=
yn∈G(vj)∩G(vi)
Wn(yn|vi)
+
yn∈G(vj)(G(vi))c
Wn(yn|vi).(68)
Observe that the second sum in the last line is bounded by
the probability Pr(Yn/∈T
δ(W|vi)|xn=vi), which in turn is
bounded by 2α1(δ)nas before, and tends to zero as well.
To bound the first sum in (68), we first consider the
cardinality of the set that the sum acts upon (the domain).
We note that since viand vjbelong to the type class T(pX)by
the first property of Lemma 6, it follows that they also belong
to the δ-typical set, i.e., vi,v
j∈T
δ(pX). Further, according to
the second property of Lemma 6, every pair of codewords vi
and vjsatisfy dH(vi,v
j). Finally, having assumed that
the rows of Waredistinct,wehavebyLemma8,
|G(vj)∩G(vi)|≤2nL(ε)|G(vj)|
2n[H(Y|X)L(ε)],(69)
where XpX, as we explained below. The second inequality
in (69) holds since the size of the conditional type class
G(xn)=Tδ(W|xn)is bounded by 2nH (Y|X)[70, Lem. 2.5],
as the type of viand vjis pX. Furthermore, by standard type
class properties [82, Th. 1.2],
Wn(yn|vi)2n[H(Y|X)δlog |Y|].(70)
Now by Equation (69) and (70),
yn∈G(vj)∩G(vi)
Wn(yn|vi)2n[L(ε)δlog |Y|],(71)
which tends to zero as n→∞for sufficiently small δ>0,
such that δlog |Y| <L(ε). Thus, by (68) and (71), the
probability of type II error is bounded by
Pe,2(i, j)22(ε,δ),(72)
for sufficiently large n,whereα2(ε, δ)=min{α1(δ),
L(ε)δlog |Y|}. The proof follows by taking the limits
n→∞,andε,δ0.
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
10 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 68, NO. 1, JANUARY 2022
D. Converse Proof
To prove the converse part, we will use the following
observation. Let R>0be an achievable rate. We will
assume to the contrary that there exist two different messages
i1and i2that are represented by the same codeword, i.e.,
ui1=ui2=xn, and show that this leads to error probabilities
such that
Pe,1(i1)+Pe,2(i2,i
1)=1.(73)
Hence the assumption is false. The number of messages 2nR
is thus bounded by the size of the subset of input sequences
that satisfy the input constraint φn(xn)A. Then we notice
that the average cost of a codeword depends only on its type,
and hence this subset is in fact a union of type classes. This
also implies that we have a strong converse for the DI capacity.
Consider a sequence of (2nR,n
(n)
1
(n)
2)codes
(U(n),D(n))such that λ(n)
1and λ(n)
2tend to zero as
n→∞.
Lemma 9: Consider a sequence of codes as described above.
Then, given a sufficiently large n, the codebook U(n)satisfies
the following property. There cannot be two distinct messages
that are represented by the same codeword, i.e.,
i1=i2ui1=ui2,(74)
where i1,i
2[[ 2 nR]] .
Proof: Assume to the contrary that there exist two mes-
sages i1and i2,wherei1=i2, such that
ui1=ui2=xn,(75)
for some xn∈X
n.Since(U(n),D(n))form a
(2nR,n
(n)
1
(n)
2)code, we have
Pe,1(i1)=Wn(Dc
i1|xn)λ(n)
1
Pe,2(i2,i
1)=Wn(Di1|xn)λ(n)
2.(76)
This leads to a contradiction as
1=Wn(Dc
i1|xn)+Wn(Di1|xn)
=Pe,1(i1)+Pe,2(i2,i
1)
λ(n)
1+λ(n)
2.(77)
Hence, the assumption is false, and i1and i2cannot have
the same codeword.
By Lemma 9, each message has a distinct codeword. Hence,
the number of messages is bounded by the number of input
sequences that satisfy the input constraint. That is, the size of
the codebook is upper-bounded as follows:
2nR xn:1
n
n
t=1
φ(xt)A.(78)
Notice that the input cost of a given sequence xndepends
only on the type of the sequence, since
1
n
n
t=1
φ(xt)=
a∈X
ˆ
Pxn(a)φ(a)
=Eφ(X),(79)
where the random variable Xis distributed according to the
type of xn, i.e., pX=ˆ
Pxn. Therefore, the subset on the right
hand side of (78) can be written as a union of type classes:
xn:1
n
n
t=1
φ(xt)A
=
pX∈Pn(X):
E{φ(X)}≤A
T(pX)
≤|P
n(X)|max
pX∈Pn(X):
E{φ(X)}≤A
|T (pX)|
≤|P
n(X)2nH(X)
2n(H(X)+αn)
2n(CDI (W)+αn),(80)
where αn0as n→∞,wherePn(X)denotes the space
of all types over Xof sequences of length n. The second
inequality holds since the size of a type class T(pX)is
bounded by |T (pX)|≤2nH(X)[81, Th. 11.1.3]. The third
inequality holds since the number of types on Xis polynomial
in n[81, Th. 11.1.1]. Thus, by (78) and (80), the code rate is
bounded by RCDI (W)+αn, which completes the proof
of Theorem 4.
IV. FADING CHANNELS
In this section we consider Gaussian channels with either
fast fading or slow fading. We will see that the capacity
characterization is inherently different in the sense that for the
Gaussian channel, the code size scales L(n, R)=2
nlog(n)R=
nnR. We note that the scale of the DI capacity can be viewed
as a special case of a tetration function, as 2n=nn=
2nlog(n)[83], [84]. To prove this property, we establish lower
and upper bounds in this scale, both positive and finite. As a
consequence, it follows that the capacity is infinite in the
exponential scale L(n, R)=2
nR and zero in the double
exponential scale L(n, R)=2
2nR .
A. Fading Channels
Consider the Gaussian channel Gfast with fast fading, spec-
ified by the input-output relation
Y=Gx+Z,(81)
where Gis a random sequence of fading coefficients and
Zis an additive white Gaussian process (see Figure 3).
Specifically, Gis a sequence of i.i.d. continuous random
variables fGwith finite moments, while the noise sequence
Zis i.i.d. ∼N(0
2
Z). It is assumed that the noise sequence
Zand the sequence of fading coefficients Gare statistically
independent, and that the values of the fading coefficients
belong to a bounded set G, either countable or uncountable.
The transmission power is limited to x2nA.
Similarly, the Gaussian channel Gslow with slow fading is
specified by the input-output relation
Yt=Gxt+Zt,(82)
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
SALARISEDDIGH et al.: DETERMINISTIC IDENTIFICATION OVER CHANNELS WITH POWER CONSTRAINTS 11
Fig. 3. Deterministic identification over fading channels. For fast fading,
G=(Gt)
t=1 is a sequence of i.i.d. fading coefficients fG.Forslow
fading, the fading sequence remains constant throughout the transmission
block, i.e., Gt=G.
where Gis a continuous random variable fG(g). Suppose
that the values of Gbelong to a set G,andthatGhas finite
expectation and finite variance var(G)>0with additive
white Gaussian noise, i.e., where the noise sequence Zis i.i.d.
∼N(0
2
Z). The transmission power is limited to x2nA.
B. Coding for Fading Channels
A code for the Gaussian channel with fast fading is defined
below.
Definition 4: An (L(n, R),n)DI code with channel side
information (CSI) at the decoder for a Gaussian channel Gfa st
under input constraint A, assuming L(n, R)is an integer,
is defined as a system (U,D)which consists of a codebook
U={ui}i[[ L(n,R)]],U⊂Rn, such that
ui2nA, for all i[[ L(n, R)]],(83)
and a collection of decoding regions D=
{Di,g}i[[ L(n,R)]],g∈Gnwith L(n,R)
i=1 Di,gRnGiven
a message i[[ L(n, R)]], the encoder transmits ui.The
decoder’s aim is to answer the following question: Was a
desired message jsent or not? There are two types of errors
that may occur: Rejecting the true message, or accepting a
false message. Those are referred to as type I and type II
errors, respectively.
The error probabilities of the identification code (U,D)are
given by
Pe,1(i)=1$Gn
fG(g)%$Di,g
fZ(ygui)dy&dg,
(84)
Pe,2(i, j)=$Gn
fG(g)%$Dj,g
fZ(ygui)dy&dg,(85)
with fZ(z)= 1
(2πσ2
Z)n/2e−z2/2σ2
Z(see Figure 3). An
(L(n, R),n
1
2)DI code further satisfies
Pe,1(i)λ1,(86)
Pe,2(i, j)λ2,(87)
for all i, j [[ L(n, R)]], such that i=j.ArateR>0is called
achievable if for every λ1
2>0and sufficiently large n,
there exists an (L(n, R),n
1
2)) DI code. The operational
DI capacity in the L-scale is defined as the supremum of
achievable rates, and will be denoted by CDI(Gfast ,L).
Coding for the Gaussian channel with slow fading is defined
as in the compound channel model, considering the worst-case
channel. Thus, the error is maximized over the set of values of
the fading coefficients. A code for the Gaussian channel with
slow fading is defined in a similar manner as in Definition 4.
However, the errors are defined with a supremum over the
values of the fading coefficient G∈G, namely,
Pe,1(i)=sup
g∈G '1$Di,g n
t=1
fZ(ytgui,t)dy(,
(88)
Pe,2(i, j)=sup
g∈G '$Dj,g n
t=1
fZ(ytgui,t)dy(.(89)
The capacity of the Gaussian channel with slow fading is
denoted by CDI (Gslow,L).
C. Main Results–Fading Channels
We determine the coding scale for the DI capacity of
Gaussian channels with fading. Before we give our results,
we make the following observation. Recall that we use the
notation of L2L1for a coding scale L1that dominates
L2(see Definition 1). The following property readily follows
from the definition. Suppose that the capacity in a given scale
is positive yet finite, i.e., 0<CDI (Gfast ,L
0)<for a given
L0. Then, for every LL0,
CDI(Gfast ,L
)=,(90)
and for every L+L0,
CDI(Gfast ,L
+)=0.(91)
Our DI capacity theorem for the Gaussian channel with fast
fading is stated below.
Theorem 10: Assume that the fading coefficients are pos-
itive and bounded away from zero, i.e., 0/cl(G).TheDI
capacity of the Gaussian channel Gfas t with fast fading in the
2nlog(n)-scale, i.e., for L(n, R)=2
(nlog n)Ris bounded by
1
4CDI(Gfast ,L)1.(92)
Hence, the DI capacity is infinite in the exponential scale
and zero in the double exponential scale, i.e.,
CDI(Gfast ,L)=for L(n, R)=2
nR,
0for L(n, R)=2
2nR .(93)
The proofs for the lower and upper bounds in the first
part of Theorem 10 are given in Subsection IV-D and
Subsection IV-E, respectively. The second part of the theo-
rem is a direct consequence of the observation given at the
beginning of this subsection (see (90)-(91)).
Remark 6: Observe that Theorem 10 assumes that the
fading coefficients are positive and bounded away from zero,
i.e., 0/cl(G). In practice, however, communication with
fading may involve small gain that can be close to zero. For
instance, if the fading distribution fGis Gaussian, then the
probability Pr(|Gt|)is positive for arbitrarily small ε>0.
Hence, Gaussian fading does not meet the assumption in our
theorem. Unfortunately, the case where the fading coefficients
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
12 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 68, NO. 1, JANUARY 2022
can be arbitrarily close (or equal) to zero remains unsolved.
We give a rough explanation of how we use the assumption
in the analysis. In the achievability proof in Subsection IV-D,
we assume that there exists γ>0such that Gtfor all
twith probability 1. The codebook is constructed such that
the codewords are distanced from each other by n. Then,
in the analysis for the type II error, we consider an error event
of the form
Z2n(σ2
Z+2δn)−G(uiuj)2,(94)
(see (120)). Given our assumption, we have
G(uiuj)2γ2uiuj2. By choosing εn=3δn
γ2,
the error event in (94) implies
Z2n(σ2
Z+2δnγ2εn)=n(σ2
Zδn),(95)
or equivalently,
1
n
i=1
Z2
iσ2
Z≤−δn.(96)
As the random sequence Z2
iis i.i.d. with E(Z2
i)=σ2
Z,
we show that this probability tends to zero using large devia-
tions arguments. Further details are given in Subsection IV-D.
Our DI capacity theorem for the Gaussian channel with slow
fading is stated below.
Theorem 11: The DI capacity of the Gaussian channel Gslow
with slow fading in the 2nlog(n)-scale is bounded by
1
4CDI(Gslow,L)1if 0/cl(G)
CDI(Gslow,L)=0 if 0cl(G),
for L(n, R)=2
nlog(n)R.(97)
Hence, the DI capacity is infinite in the exponential scale,
if 0/cl(G), and zero in the double exponential scale, i.e.,
CDI(Gslow,L)=0if 0cl(G)
if 0/cl(G),
for L(n, R)=2
nR,(98)
CDI (Gslow,L)=0,for L(n, R)=2
2nR .(99)
The derivation of the result above is similar to that of
the proof for fast fading in Section IV-C. For completeness,
we give the proof of Theorem 11 in Appendix B.
D. Lower Bound (Achievability Proof for Theorem 10)
Consider the Gaussian channel Gfast with fast fading.
We show that the DI capacity is bounded by CDI (Gfast ,L)
1
4for L(n, R)=2
nlog(n)R. Achievability is established using
a dense packing arrangement and a simple distance-decoder.
A DI code for the Gaussian channel Gfast with fast fading is
constructed as follows. Consider the normalized input-output
relation,
¯
Y=G¯
x+¯
Z,(100)
where the noise sequence ¯
Zis i.i.d. ∼N0,σ2
Z
n,andan
input power constraint
¯
x≤A, (101)
Fig. 4. Illustration of a sphere packing, where small spheres of radius
r0=εncover a bigger sphere of radius r1=Aεn.Thesmall
spheres are disjoint from each other and have a non-empty intersection with
the big sphere. Some of the small spheres, marked in gray, are not entirely
contained within the bigger sphere, and yet they are considered to be a part
of the packing arrangement. As we assign a codeword to each small sphere
center, the norm of a codeword is bounded by Aas required.
with ¯
x=1
nx,¯
Z=1
nZ,and ¯
Y=1
nY. Assuming
0/cl(G), there exists a positive number γsuch that
|Gt|, (102)
for all twith probability 1.
1) Codebook Construction: We use a packing arrangement
of non-overlapping hyper-spheres of radius εnin a hyper-
sphere of radius (Aεn), with
εn=A
n1
2(1b),(103)
where b>0is arbitrarily small.
Let Sdenote a sphere packing, i.e., an arrangement of L
non-overlapping spheres Sui(n, r0),i[[ L(n, R)]] that cover a
bigger sphere S0(n, r1), with r1>r
0. As opposed to standard
sphere packing coding techniques, the small spheres are not
necessarily entirely contained within the bigger sphere. That
is, we only require that the spheres are disjoint from each
other and have a non-empty intersection with S0(n, r1).See
illustration in Figure 4. The packing density Δn(S)is defined
as the fraction of the big sphere volume Vol (S0(n, r1)) that
is covered by the small spheres, i.e.
Δn(S)Vo l S0(n, r1)L
i=1 Sui(n, r0)
Vo l (S0(n, r1)) ,(104)
[85, see Ch. 1]. A sphere packing is called saturated if no
spheres can be added to the arrangement without overlap.
We use a packing argument that has a similar flavor as in the
Minkowski–Hlawka theorem in lattice theory [85]. We use the
property that there exists an arrangement L
i=1 Sui(n, εn)of
non-overlapping spheres inside S0(n, A)with a density of
Δn(S)2n[78, Lem. 2.1]. Specifically, consider a satu-
rated packing arrangement of L(n, R)=2
nlog(n)Rspheres
of radius r0=εncovering the big sphere S0(n, r1=
Aεn), i.e., such that no spheres can be added without
overlap. Then, for such an arrangement, there cannot be a
point in the big sphere S0(n, r1)with a distance of more than
2r0from all sphere centers. Otherwise, a new sphere could
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
SALARISEDDIGH et al.: DETERMINISTIC IDENTIFICATION OVER CHANNELS WITH POWER CONSTRAINTS 13
be added. As a consequence, if we double the radius of each
sphere, the 2r0-radius spheres cover the whole sphere of radius
r1. In general, the volume of a hyper-sphere of radius ris
given by
Vo l (Sx(n, r)) = πn
2
Γ(n
2+1) ·rn,(105)
(see Eq. (16) in [85]). Hence, doubling the radius multiplies
the volume by 2n. Since the 2r0-radius spheres cover the
entire sphere of radius r1, it follows that the original r0-radius
packing has a density of at least 2n, i.e.,
Δn(S)2n.(106)
We assign a codeword to the center uiof each small
sphere. The codewords satisfy the input constraint as
ui≤r0+r1=A. Since the small spheres have the
same volume, the total number of spheres is bounded from
below by
L=
Vo l L
i=1 Sui(n, r0)
Vo l (Su1(n, r0))
Vo l S0(n, r1)L
i=1 Sui(n, r0)
Vo l (Su1(n, r0))
=Δn(S)·Vo l (S0(n, r1)))
Vo l (Su1(n, r0))
2n·Vo l (S0(n, r1)))
Vo l (Su1(n, r0))
=2
n·rn
1
rn
0
,(107)
where the second equality is due to (104), the inequality
that follows holds by (106), and the last equality follows
from (105). That is, the codebook size satisfies
L(n, R)=2
nlog(n)R
2n·Aεn
εnn
.(108)
Hence,
R1
log(n)log Aεn
εn1
log(n)
=1
log(n)log n1
4(1b)11
log(n)
1
log(n)log n1
4(1b)11
log(n)
=1
4(1 b)2
log(n),(109)
which tends to 1
4when n→∞and b0, where the second
inequality holds, since log(t1) log(t)1for t2.
2) Encoding: Given a message i[[ L(n, R)]],
transmit ¯
x=¯
ui.
3) Decoding: Let
δn=γ2εn
3=γ2A
3n1
2(1b).(110)
To identify whether a message j[[ L(n, R)]] was sent,
given the sequence g, the decoder checks whether the channel
output ¯
ybelongs to the following decoding set,
Dj,g=¯
yRn:¯
yg¯
uj≤)σ2
Z+δn.(111)
4) Error Analysis: Consider the type I error, i.e., when the
transmitter sends ¯
ui,yet ¯
Y/∈D
i,G.Foreveryi[[ L(n, R)]],
the type I error probability is bounded by
Pe,1(i)=Pr*
*¯
YG¯
ui*
*2
2
Z+δn¯
x=¯
ui
=Pr*
*¯
Z*
*2
2
Z+δn
=Prn
t=1
¯
Zt
2
2
Z+δn
Pr n
t=1
¯
Zt
2
2
Z+δn
3σ4
Z
2
n
=27σ4
Z
nbA2γ4
λ1,(112)
for sufficiently large nand arbitrarily small λ1>0,where
the second inequality follows by Chebyshev’s inequality, and
since the fourth moment of a Gaussian variable V∼N(0
2
V)
is E{V4}=3σ4
V.
Next we address the type II error, i.e., when ¯
Y∈D
j,G
while the transmitter sent ¯
ui. Then, for every i, j [[ L(n, R)]],
where i=j, the type II error probability is given by
Pe,2(i, j)=Pr*
*¯
YG¯
uj*
*2σ2
Z+δn¯
x=¯
ui
=Pr*
*G(¯
ui¯
uj)+¯
Z*
*2σ2
Z+δn.
(113)
Observe that the square norm can be expressed as
*
*G(¯
ui¯
uj)+¯
Z*
*2=G(¯
ui¯
uj)2+*
*¯
Z*
*2
+2
n
t=1
Gtui,t ¯uj,t)¯
Zt.(114)
Then, define the event
E0=
n
t=1
Gtui,t ¯uj,t )¯
Zt>δn
2.(115)
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
14 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 68, NO. 1, JANUARY 2022
By Chebyshev’s inequality, the probability of this event
vanishes,
Pr(E0)n
t=1ui,t ¯uj,t)2E{G2
t}E{¯
Z2
t}
δn
22
=σ2
Z(σ2
G+μ2
G)n
t=1ui,t ¯uj,t)2
nδn
22
=4σ2
Z(σ2
G+μ2
G)¯
ui¯
uj2
2
n
,(116)
where the first inequality holds since the sequences {¯
Zt}and
{Gt}are i.i.d. ∼N0,σ2
Z
nand fGwith E{Gt}=μG
and E{G2
t}=σ2
G+μ2
G. By the triangle inequality,
¯
ui¯
uj2(¯
ui+¯
uj)2
(A+A)2
=4A, (117)
hence
Pr(E0)162
Z(σ2
G+μ2
G)
2
n
=144σ2
Z(σ2
G+μ2
G)
γ4Anb
η1,(118)
for sufficiently large n,whereη1>0is arbitrarily small.
Furthermore, observe that given the complementary event Ec
0,
we have 2n
t=1 Gtui,t ¯uj,t )¯
Zt≥−δn, hence
*
*G(¯
ui¯
uj)+¯
Z*
*2σ2
Z+δn
G(¯
ui¯
uj)2+*
*¯
Z*
*2σ2
Z+2δn,(119)
(see (114)). Therefore, applying the law of total probability,
we have
Pe,2(i, j)
(a)
=Pr
*
*G(¯
ui¯
uj)+¯
Z*
*2σ2
Z+δnand
n
t=1
Gtui,t ¯uj,t)¯
Zt>δn
2
+Pr*
*G(¯
ui¯
uj)+¯
Z*
*2σ2
Z+δnand
n
t=1
Gtui,t ¯uj,t)¯
Ztδn
2
(b)
Pr(E0)+Pr*
*G(¯
ui¯
uj)+¯
Z*
*2σ2
Z+δnand
2
n
t=1
Gtui,t ¯uj,t)¯
Zt≥−δn
(c)
Pr(E0)+PrG(¯
ui¯
uj)2+*
*¯
Z*
*2σ2
Z+2δn
(d)
Pr G(¯
ui¯
uj)2+*
*¯
Z*
*2σ2
Z+2δn+η1,
(120)
where (a)follows by applying the law of total probability
to (113), (b)follows from (115), (c)holds by (119), and (d)
by (118).
Based on the codebook construction, each codeword is
surrounded by a sphere of radius εn, which implies that
¯
ui¯
uj≥εn.(121)
Then, by (102),
G(¯
ui¯
uj)2γ2¯
ui¯
uj2γ2εn,(122)
where γis the minimal value in G. Hence, according to (120),
Pe,2(i, j)Pr *
*¯
Z*
*2σ2
Z+2δnγ2εn+η1
Pr *
*¯
Z*
*2σ2
Zδn+η1,(123)
where the last line holds, since 2δnγ2εn=δnby (110).
Therefore, by Chebyshev’s inequality,
Pe,2(i, j)Pr n
t=1
¯
Z2
tσ2
Z≤−δn+η1
n
t=1 var(¯
Z2
t)
δ2
n
+η1
n
t=1 E{¯
Z4
t}
δ2
n
+η1
=
n·3σ2
Z
n2
δ2
n
+η1
=27σ4
Z
γ4A2nb+η1
λ2,(124)
for sufficiently large n,whereλ2is arbitrarily small.
We have thus shown that for every λ1
2>0and
sufficiently large n, there exists a (2nlog(n)R,n
1
2)code.
As we take the limits of n→∞,andthenb0,the
lower bound on the achievable rate tends to 1
4, by (109). This
completes the achievability proof for Theorem 10.
E. Upper Bound (Converse Proof for Theorem 10)
We show that the capacity is bounded by CDI (Gfas t ,L)1.
We note that in the converse proof, we do not normalize the
sequences. Suppose that Ris an achievable rate in the L-scale
for the Gaussian channel with fast fading. Consider a sequence
of (L(n, R),n
(n)
1
(n)
2)codes (U(n),D(n))such that λ(n)
1
and λ(n)
2tend to zero as n→∞.
We begin with the following lemma.
Lemma 12: Consider a sequence of codes as described
above. Let b>0be an arbitrarily small constant that does
not depend on n. Then there exists n0(b), such that for all
n>n
0(b), every pair of codewords in the codebook U(n)are
distanced by at least n,i.e.,
ui1ui2≥n,(125a)
where
εn=A
n2(1+b),(125b)
for all i1,i
2[[ L(n, R)]] such that i1=i2.
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
SALARISEDDIGH et al.: DETERMINISTIC IDENTIFICATION OVER CHANNELS WITH POWER CONSTRAINTS 15
Proof: Fix λ1and λ2.Letκ, θ, ζ > 0be arbitrarily small.
Assume to the contrary that there exist two messages i1and
i2,wherei1=i2, such that
ui1ui2<n=αn,(126)
where
αn=A
n1
2(1+2b).(127)
Observe that
EG(ui1ui2)2=
n
t=1
EG2
t(ui1,t ui2,t)2
=EG2ui1ui22,(128)
and consider the subsets
Ai1,i2={g∈G
n:g(ui1ui2)
n},(129)
Fi2=yRn:ygui,2≤)n(σ2
Z+ζ),
(130)
where
δn=A
n1
2(1+b).(131)
By Markov’s inequality, the probability that the fading
sequence Gbelongs to this set is bounded by
Pr (G∈A
i1,i2)=Pr(G(ui1ui2)2
2
n)
(a)
E{G2}ui1ui22
δ2
n
(b)
E{G2}α2
n
δ2
n
=E{G2}
nb
κ, (132)
for sufficiently large nwhere (a)holds since the sequence
{Gt}n
t=1 is i.i.d. and (b)is due to (126).
Then, observe that
1Pe,1(i1)
=$Gn
fG(g)%$Di1,g
fZ(ygui1)dy&dg
$Ac
i1,i2
fG(g)%$Di1,g
fZ(ygui1)dy&dg
+Pr(G∈A
i1,i2)
$Ac
i1,i2
fG(g)%$Di1,g
fZ(ygui1)dy&dg+κ. (133)
Hence,
1κPe,1(i1)
$Ac
i1,i2
fG(g)%$Di1,g
fZ(ygui1)dy&dg
=$Ac
i1,i2
fG(g)%$Di1,g∩Fi2
fZ(ygui1)dy
+$Di1,g∩Fc
i2
fZ(ygui1)dy&dg
$Ac
i1,i2
fG(g)%$Fi2
fZ(ygui1)dy
+$Fc
i2
fZ(ygui1)dy&dg.(134)
Consider the second integral, where ygui,2>
+n(σ2
Z+ζ)(see (130)). Then, by the triangle inequality
ygui,1≥ygui,2−g(ui,1ui,2)
>)n(σ2
Z+ζ)−g(ui,1ui,2)
)n(σ2
Z+ζ)δn,(135)
for every g∈A
c
i1,i2(see (126)). For sufficiently large n,this
implies
ygui,1≥)n(σ2
Z+η),(136)
for η<ζ
2.Thatis,Fi2implies ygui,1≥
+n(σ2
Z+η).
We deduce that for every g∈A
c
i1,i2, the subset of output
sequences y∈Ynsuch that ygui,2>+n(σ2
Z+ζ)is
included within the subset of sequences with ygui,1>
+n(σ2
Z+η). Hence, the second integral in the right hand side
of (134) is bounded by
$
y:ygui,1>n(σ2
Z+η)
fZ(ygui1)dy
=$
z:z>n(σ2
Z+η)
fZ(z)dz
=Pr(Z22
Z>nη)
3σ4
Z
2
κ, (137)
for large n, where the first inequality is due to Chebyshev’s
inequality, followed by the substitution of zygui1.
Thus, by (134),
12κPe,1(i1)
$Ac
i1,i2
fG(g)%$Di1,g∩Fi2
fZ(ygui1)dy&dg.(138)
Now, we can focus on g∈A
c
i1,i2and y∈F
i2, such that
ygui,2≤)n(σ2
Z+ζ).(139)
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
16 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 68, NO. 1, JANUARY 2022
Observe that
fZ(ygui1)fZ(ygui2)
=fZ(ygui1)%1e1
2σ2
Z
ygui22ygui12
&.
(140)
By the triangle inequality,
ygui1≤ygui2+g(ui1ui2).(141)
Taking the square of both sides, we have
ygui12
≤ygui22+g(ui2ui1)2
+2ygui2·g(ui2ui1)
≤ygui22+δ2
n+2δn)n(σ2
Z+ζ)
=ygui22+δ2
n+2+A(σ2
Z+ζ)
nb
2
,(142)
where the last inequality follows from the definition of Ai1,i2
in (129), and due to (139). Thus, for sufficiently large n,
ygui12−ygui22θ. (143)
Hence,
fZ(ygui1)fZ(ygui2)
fZ(ygui1)(1 e1
2σ2
Z
θ)
κfZ(ygui1),(144)
for sufficiently small θ>0, such that 1e1
2σ2
Z
θκ.Now
by (138), we get
λ1+λ2
Pe,1(i1)+Pe,2(i2,i
1)
12κ
$Ac
i1,i2
fG(g)%$Di1,g∩Fi2
fZ(ygui1)dy&dg
+$Gn
fG(g)%$Di1,g
fZ(ygui2)dy&dg
12κ
$Ac
i1,i2
fG(g)%$Di1,g∩Fi2
fZ(ygui1)dy&dg
+$Ac
i1,i2
fG(g)%$Di1,g∩Fi2
fZ(ygui2)dy&dg
12κ$Ac
i1,i2
fG(g)%$Di1,g∩Fi2
(fZ(ygui1)
fZ(ygui2))dy&dg.(145)
Hence, by (144),
λ1+λ212κ
κ$Ac
i1,i2
fG(g)%$Di1,g∩Fi2
fZ(ygui1)dy&dg
13κ, (146)
which leads to a contradiction for sufficiently small κsuch that
3κ<1λ1λ2. This completes the proof of Lemma 12.
By Lemma 12, we can define an arrangement of non-
overlapping spheres Sui(n, n)of radius ncentered
at the codewords ui. Since the codewords all belong to a
sphere S0(n, nA)of radius nA centered at the origin,
it follows that the number of packed spheres, i.e., the number
of codewords 2nlog(n)R, is bounded by
2nlog(n)RVo l (S0(n, nA +n))
Vo l (Sui(n, n))
=A+εn
εnn
.(147)
Thus,
R1
log nlog A+εn
εn
=log 1+n1+b
log n
=log n1+b1+ 1
n1+b
log n
=(1 + b)logn+log1+ 1
n1+b
log n
=1+b+log 1+ 1
n1+b
log n,(148)
which tends to 1+bas n→∞. Therefore, R1+b.Now,
since b>0is arbitrarily small, an achievable rate must satisfy
R1. This completes the proof of Theorem 10.
V. S UMMARY AND DISCUSSION
We have established the deterministic identification (DI)
capacity of a DMC subject to an input constraint and have
developed lower and upper bounds on the DI capacity of
Gaussian channels with fast fading and slow fading in the
scale L(n, R)=2
nlog(n)R=nnR,wherenis the blocklength.
We have thus established that the super-exponential scale nnR
is the appropriate scale for the DI capacity of the Gaussian
channel. That is, the DI capacity can only be positive and finite
in this coding scale. This scale is sharply different from the
usual scales in the transmission and randomized-identification
settings, where the code size scales exponentially and double
exponentially, respectively. Different non-standard scales are
also observed in other communication models, such as covert
identification [86], [87], where the identification message is of
size 22nR . For the DMC, the DI capacity formula is given
in terms of the entropy of the reduced channel input (see
Definition 3), and does not depend on the specific transition
probabilities corresponding to the reduced input alphabet.
We have mentioned molecular communication as a moti-
vating application of this study. Recently, there have been
significant advances in molecular communication for complex
nano-networks. The interconnection of nanothings with the
Internet is known as the Internet of NanoThings (IoNT)
and is the basis for various future healthcare and military
applications [88]. Furthermore, the concept of the Internet of
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
SALARISEDDIGH et al.: DETERMINISTIC IDENTIFICATION OVER CHANNELS WITH POWER CONSTRAINTS 17
Bio-NanoThings (IoBNT) has been introduced in [89], where
nanothings are biological cells that are created using tools from
synthetic biology and nanotechnology. For the communication
between cells, molecular communication is well suited, since
the natural exchange of information between cells is already
based on this paradigm. Molecular communication in cells is
based on signal pathways (chains of chemical reactions) that
process information that is modulated into chemical charac-
teristics, such as molecule concentration. Identification is of
interest for these applications. In the case of Bio-NanoThings,
it is uncertain whether natural biological processes can be
controlled or reinforced by local randomness at this level.
Therefore, for the design of synthetic IoNT, or for the analysis
and utilization of IoBNT, identification with deterministic
encoding is more appropriate.
Gaussian channels have an entirely different and unusual
behaviour. We have considered deterministic identification for
Gaussian channels with fast fading and slow fading, where CSI
is available at the decoder. The DI capacity of the standard
memoryless Gaussian channel Y=X+Zis infinite in the
exponential scale, i.e.,
CDI(G,L)=,for L(n, R)=2
nR,(149)
as we have recently observed in [2] and [77]. However, for
a finite blocklength n, the number of codewords must be
finite. Thereby, the meaning of the infinite capacity result
is that the number of messages scales super-exponentially.
This raises the question: What is the true order of the code
size. In mathematical terms, what is the scale L(n, R)for
which the DI capacity is positive yet finite. To answer this
question, we have shown that the number of messages scales
as 2nlog(n)R. As a consequence, we have deduced that the
DI capacity of a Gaussian channel with fast fading is infinite
in the exponential scale, and zero in the double exponential
scale, regardless of the channel noise. For slow fading, the DI
capacity in the exponential scale is infinite, unless the fading
gain can be zero or arbitrarily close to zero (with positive
probability), in which case the DI capacity is zero. Note,
however, that this scale is comparably lower than the double
exponential scale of RI coding.
Next, we compare and discuss different results from the
literature on the DI capacity. For the double exponential scale,
or equivalently, when the rate is defined as R=1
nlog log (#
of messages), the DI capacity is
CDI (W,L)=0; for L(n, R)=2
2nR ,(150)
since the code size of DI codes scales only exponentially
in block length. On the other hand, as observed by Bracher
and Lapidoth [33], if one considers an average error criterion
instead of the maximal error, then the double exponential per-
formance of randomized-encoder codes can also be achieved
using deterministic codes.
By providing a detailed proof for the DI capacity theorem
with and without an input constraint, we have filled the
gap in the previous analysis [4], [43] as well. In particular,
in [43], Ahlswede and Cai asserted that the DI capacity for a
compound channel is given by
CDI(Wcompound,L)=max
pX
min
sH(ˆ
X(s)) ;
for L(n, R)=2
nR,(151)
where s∈Sis the channel state, and the map ˆ
X(s)is induced
from Xby a partition of the input alphabet to equivalent
classes as specified in [43, Sec. I.F]. This result immediately
yields Corollary 5, since the DMC is a special case of a
compound channel with a single state value. Indeed, taking
|S| =1 and considering the reduced channel Wr(see Defini-
tion 3), it can be readily shown that ˆ
X(s)=X. Nonetheless,
a significant part of the proof in [43] is missing. At the
beginning of Sec. VII in [43], the following claim is given: “It
was shown in [A’80] that for any channel ˜
V:X→Ywithout
two identical rows, any u1,u
2> 0, sufficiently large nand
any U⊂X
nsuch that for all u, u∈U,d
H(u, u)>nε,
there exists a family of subsets of Yn,sayDu,u ∈U,such
that ˜
Vn(Du|u)>1u1and ˜
Vn(Du|u)<u
2for all u=u,
where dHis the Hamming distance.”, where [A’80] refers to
a paper by Ahlswede [48] on the arbitrarily varying channel,
and does not include identification.
Alternatively, one may consider the ε-capacity for a fixed
0<ε<1.ArateRis called ε-achievable in an
L-scale if there exists an (L(n, R),n)code for sufficiently
large n(see Definition 2). The DI ε-capacity Cε
DI(W,L)is
then defined as the supremum of ε-achievable rates. As the
RI capacity in the double exponential scale has a strong
converse [4], [80], [90], for L(n, R)=Ldouble(n, R)and
0<ε<1
2,
Cε
RI (W,L
double)=CRI (W,L
double)=max
pX
I(X;Y).(152)
Based on Subsection III-D, a strong converse holds for the
DI capacity as well, hence
Cε
DI(W,L
double)=CDI (W,L
double)=0.(153)
On the other hand, for ε1
2
Cε
DI(W,L
double)=Cε
RI (W,L
double)=.(154)
To understand (154), suppose ε> 1
2, and consider an
arbitrary set of codewords with a stochastic decoder that makes
a decision for the identification hypothesis by flipping a fair
coin [4]. Both error probabilities of type I and II equal 1
2,
and are thus smaller than ε. Similarly, for the Gaussian
channel, [30], [39],
Cε
RI (G,L
double)=CRI (G,L
double)= 1
2log 1+ A
σ2
Z,
for 0<ε<1
2,(155)
Cε
DI(G,L
double)=Cε
RI (G,L
double)=,for ε1
2.
(156)
Based on our result in Theorem 10, with Gt1for all t,
we have that
1
4CDI(G,L
)Cε
DI(G,L
)1,(157)
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
18 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 68, NO. 1, JANUARY 2022
with
L(n, R)=2
nlog(n)R=nnR.(158)
Hence, in the double exponential scale,
Cε
DI(G,L
double)=CDI (G,L
double)=0.(159)
APPENDIX A
PROOF OF LEMMA 3
Let Wbe a given DMC, with a stochastic matrix W:
X→Yand its reduced version Wr:Xr→Yas defined in
Definition 3. Observe that the capacity of the original channel
is lower bounded by that of the reduced channel, i.e.,
CDI (W,L)CDI(Wr,L),(160)
since every code for Wrcan also be used for W. Hence,
it remains to be shown that
CDI (Wr,L)CDI(W,L).(161)
Assume without loss of generality that the input alphabet
of the original channel Wis given by X={1,2,···,|X|}.
Let σ:X→X
rdenote the projection of the input alphabet
onto the equivalent classes,
σ[x]=z()iff x∈X().(162)
Now let (U,D)be an (L(n, R),n
1
2)code for W.Then
the type I probability of error can be expressed as
PW
e,1(i)=
yn/∈Di
Wn(yn|ui)=
yn/∈Di
n
t=1
W(y(t)|ui(t)),
(163)
where we use the notation yn=y(t)n
t=1 and ui=
ui(t)n
t=1.Nextwedeneacode(,
U,D)for the channel Wr,
where the codebook consists of the following codewords,
˜ui=σ[ui(t)]n
t=1.(164)
Now recall that we have defined the equivalence classes
such that input letters in the same equivalence class correspond
to identical rows in the channel matrix W(see Definition 3).
Thus, by definition,
Wr(y|σ[x]) = W(y|x),(165)
for all x∈Xand y∈Y. Hence, the error probability of type
I for the reduced channel Wrsatisfies
PWr
e,1(i)(a)
=
yn/∈Di
Wr(yn|˜ui)
(b)
=
yn/∈Di
n
t=1
Wry(t)|σ[ui(t)]
(c)
=
yn/∈Di
n
t=1
W(y(t)|ui(t))
(d)
=
yn/∈Di
Wn(yn|ui)
(e)
=PW
e,1(i),(166)
for all i,where(a)and (e)are due to (9); (b)and (d)hold
since the channel is memoryless, and (c)follows from (165).
By the same considerations, we also have PWr
e,2(i, j)=
PW
e,2(i, j)for all j=i. That is, the error probabilities of
the code (,
U,D)are the same as those of the original code
for W. Therefore, the code constructed above for Wris
also an (L(n, R),n
1
2)code, and the proof of Lemma 3
follows.
APPENDIX B
PROOF OF THEOREM 11
Consider the Gaussian channel Gslow with slow fading.
When 0cl(G), it immediately follows that the DI capacity is
zero. To see this, observe that if 0cl(G), then by (88)-(89),
Pe,1(i)+Pe,2(j, i)
=sup
g∈G '1$Di,g n
t=1
fZ(ytgui,t)dy(
+sup
g∈G '$Dj,g n
t=1
fZ(ytgui,t)dy(
'1$Di,g n
t=1
fZ(ytgui,t)dy(g=0
+'$Di,g n
t=1
fZ(ytguj,t)dy(g=0
=1.(167)
Hence, in the sequel we suppose that 0/cl(G).
A. Lower Bound (Achievability Proof)
We show here that the DI capacity of the Gaussian channel
with slow fading can be achieved using a dense packing
arrangement and a simple distance-decoder. A DI code for
the Gaussian channel Gslow with slow fading is constructed as
follows. Since the decoder can normalize the output symbols
by 1
n, we have an equivalent input-output relation,
¯
Yt=G¯xt+¯
Zt,(168)
where GfG, and the noise sequence ¯
Zis i.i.d.
∼N0,σ2
Z
n, under an input power constraint
¯
x≤A, (169)
with ¯
x=1
nx,¯
Z=1
nZ,and ¯
Y=1
nY.
1) Codebook Construction: As in our achievability proof
for the fast fading setting (see Subsection IV-D.1), we use
a packing arrangement of non-overlapping hyper-spheres of
radius εnover a hyper-sphere of radius (Aεn), with
εn=A
n1
2(1b).(170)
As observed in Subsection IV-D, there exists an arrangement
2nlog(n)R
i=1 Sui(n, εn),
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
SALARISEDDIGH et al.: DETERMINISTIC IDENTIFICATION OVER CHANNELS WITH POWER CONSTRAINTS 19
over S0(n, Aεn)with a density of Δn2n[78,
Lem. 2.1]. We assign a codeword to the center of each small
sphere ui. Since the small spheres have the same volume, the
total number of spheres, i.e., the codebook size, satisfies
2nlog(n)R=
Vo l 2nlog(n)R
i=1 Sui(n, εn)
Vo l (Su1(n, εn))
2n·Vo l (S0(n, Aεn))
Vo l (Su1(n, εn))
=2
n·Aεn
εnn
,(171)
in a similar manner as in Subsection IV-D.1, hence,
R1
4(1 b)2
log(n),(172)
which tends to 1
4when n→∞and b0.
2) Encoding: Given a message i[[ L(n, R)]],transmit
¯
x=¯
ui.
3) Decoding: Let
δn=γ2εn
3=2
3n1
2(1b).(173)
To identify whether a message j[[ L(n, R)]] was sent,
given the fading coefficient g, the decoder checks whether the
channel output ¯
ybelongs to the following decoding set,
Dj,g =¯
yRn:
n
t=1
ytg¯uj,t)2)σ2
Z+δn.
(174)
4) Error Analysis: Consider the type I error, i.e., when the
transmitter sends ¯
ui,yet ¯
Y/∈D
i,G.Foreveryi[[ L(n, R)]],
the type I error probability is bounded by
Pe,1(i)=sup
g∈G
Pr n
t=1
(¯
YtG¯ui,t)2
2
Z+δn¯
x=(¯ui,t)n
t=1,G=g
=Prn
t=1
¯
Zt
2σ2
Z
n
3σ4
Z
2
n
=27σ4
Z
A2γ4nb
λ1,(175)
for sufficiently large nand arbitrarily small λ1>0,where
the second equality holds since Gand ¯
Zare statistically
independent, and the first inequality holds by Chebyshev’s
inequality and since the fourth moment of a Gaussian variable
V∼N(0
2
V)is E{V4}=3σ4
V.
Next we address the type II error, i.e., when ¯
Y∈D
j,G
while the transmitter sent ¯
ui. Then, for every i, j [[ L(n, R)]],
where i=j, the type II error probability is given by
Pe,2(i, j)=sup
g∈G -Pe,2i, j g.,(176)
where we have defined
Pe,2i, j g
Pr n
t=1
(¯
YtG¯uj,t)2σ2
Z+δn¯
x=(¯ui,t )n
t=1,G =g
=Prn
t=1
(gui,t ¯uj,t)+ ¯
Zt)2σ2
Z+δn(177)
for g∈G, as the fading coefficient Gand the noise vector
¯
Zare statistically independent. Now we bound the probability
within the square brackets.
We divide into two cases. First, consider g∈Gsuch
that g(¯
ui¯
uj)>2+σ2
Z+δn. Therefore, by the reverse
triangle inequality, ab≥|a−b|,wehave
/
0
0
1
n
t=1 (gui,t ¯uj,t)) + ¯
Zt2≥g(¯
ui¯
uj)−*
*¯
Z*
*
2)σ2
Z+δn*
*¯
Z*
*.
(178)
Hence, for every gsuch that g(¯
ui¯
uj)>2+σ2
Z+δn,
we can bound the type II error probability by
Pe,2i, j gPr *
*¯
Z*
*)σ2
Z+δn
=Prn
t=1
¯
Zt
2
2
Z+δn
3σ4
Z
2
n
=27σ4
Z
nbA2γ4
λ2,(179)
for sufficiently large nand arbitrarily small λ2>0,wherethe
second inequality follows by Chebyshev inequality.
Now, we turn to the second case, i.e., when
g(¯
ui¯
uj)≤2)σ2
Z+δn.(180)
Observe that for every given g∈G,
n
t=1
(gui,t ¯uj,t )+ ¯
Zt)2
=
n
t=1
g2ui,t ¯uj,t )2+
n
t=1
¯
Z2
t+2
n
t=1
gui,t ¯uj,t )Zt.
(181)
Then define the event
E0(g)=
n
t=1
gui,t ¯uj,t)¯
Zt>δn
2.(182)
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
20 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 68, NO. 1, JANUARY 2022
By Chebyshev’s inequality, the probability of this event
vanishes,
Pr (E0(g)) g2n
t=1ui,t ¯uj,t)2E{¯
Z2
t}
δn
22
=4σ2
Zg(¯
ui¯
uj)2
2
n
16σ2
Zσ2
Z+δn
2
n
τ0,(183)
for sufficiently large nand arbitrarily small τ0>0,where
the first inequality holds since the sequence {¯
Zt}is i.i.d.
∼N0,σ2
Z
n, and the second inequality follows from (180).
Furthermore, observe that given the complementary event
Ec
0(g),wehave2n
t=1 gui,t ¯uj,t)¯
Zt≥−δn, hence
n
t=1
(gui,t ¯uj,t )+ ¯
Zt)2σ2
Z+δn
n
t=1
g2ui,t ¯uj,t)2+
n
t=1
¯
Z2
tσ2
Z+2δn,(184)
(see (181)). Therefore, applying the law of total probability,
we have
Pe,2i, j g
=Prn
t=1
(gui,t ¯uj,t)+ ¯
Zt)2σ2
Z+δnand
n
t=1
gui,t ¯uj,t )¯
Zt>δn
2
+Prn
t=1
(gui,t ¯uj,t )+ ¯
Zt)2σ2
Z+δnand
n
t=1
gui,t ¯uj,t )¯
Ztδn
2
(a)
Pr(E0(g))
+Prn
t=1
(gui,t ¯uj,t )+ ¯
Zt)2σ2
Z+δnand
2
n
t=1
Gui,t ¯uj,t)¯
Zt≥−δn
(b)
Pr(E0(g))
+Prn
t=1
g2ui,t ¯uj,t)2+
n
t=1
¯
Z2
tσ2
Z+2δn
(c)
Pr g2¯
ui¯
uj2+*
*¯
Z*
*2σ2
Z+2δn+τ0,(185)
where (a)follows from (182), (b)holds by (184), and (c)
by (183). Based on the codebook construction, each codeword
is surrounded by a sphere of radius εn, which implies that
¯
ui¯
uj≥εn.(186)
Thus, we have
g2¯
ui¯
uj2γ2εn,(187)
where γis the minimal value in G. Hence, according to (185),
Pe,2i, j gPr *
*¯
Z*
*2σ2
Z+2δnγ2εn+τ0
=Pr*
*¯
Z*
*2σ2
Z≤−δn+τ0,(188)
(see (173)). Therefore, by Chebyshev’s inequality,
Pe,2i, j gPr n
t=1
¯
Z2
tσ2
Z≤−δn+τ0
n
t=1 var(¯
Z2
t)
δ2
n
+τ0
n
t=1 E{¯
Z4
t}
δ2
n
+τ0
=
n·3σ2
Z
n2
δ2
n
+τ0
=3σ4
Z
2
n
+τ0
=27σ4
Z
nbA2γ4+τ0
λ2,(189)
for sufficiently large n. Based on (179) and (189), we have
Pe,2i, j gλ2for all g∈G. Hence, the type II error
probability satisfies Pe,2(i, j )λ2(see (176)).
We have thus shown that for every λ1
2>0and
sufficiently large n, there exists a (2nlog(n)R,n
1
2)code.
As we take the limits of n→∞,andthenb0,the
lower bound on the achievable rate tends to 1
4, by (172). This
completes the achievability proof for Theorem 11.
B. Upper Bound (Converse Proof)
Suppose that Ris an achievable rate in the L-scale for the
Gaussian channel with slow fading. Consider a sequence of
(L(n, R),n
(n)
1
(n)
2)codes (U(n),D(n)), such that λ(n)
1and
λ(n)
2tend to zero as n→∞. We begin with the following
lemma.
Lemma 13: Consider a sequence of codes as described
above. Let b>0be an arbitrarily small constant that does
not depend on n. Then there exists n0(b), such that for all
n>n
0(b), every pair of codewords in the codebook U(n)are
distanced by at least n,i.e.,
ui1ui2≥n,(190a)
where
εn=A
n2(1+b),(190b)
for all i1,i
2[[ L(n, R)]], such that i1=i2.
Proof: Fix λ1and λ2.Letκ, θ, ζ > 0be arbitrarily small.
Assume to the contrary that there exist two messages i1and
i2,wherei1=i2, such that
ui1ui2<n=αn,(191)
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
SALARISEDDIGH et al.: DETERMINISTIC IDENTIFICATION OVER CHANNELS WITH POWER CONSTRAINTS 21
where
αnA
n1
2(1+2b).(192)
Observe that for every g∈G,
$Di1,g n
t=1
fZ(ytgui1,t)dy
=$y∈Di1,g :
n
t=1
(ytgui2,t)2n(σ2
Z+ζ)n
t=1
fZ(ytgui1,t)dy
+$y∈Di1,g :
n
t=1
(ytgui2,t)2>n(σ2
Z+ζ)n
t=1
fZ(ytgui1,t)dy
$n
t=1
(ytgui2,t)2n(σ2
Z+ζ)n
t=1
fZ(ytgui1,t)dy
+$n
t=1
(ytgui2,t)2>n(σ2
Z+ζ)n
t=1
fZ(ytgui1,t)dy.
(193)
Consider the second integral, for which ygui,2>
+n(σ2
Z+ζ), where we denote
g(g,g,...,g).(194)
Then, by the triangle inequality,
ygui,1≥ygui,2−g(ui,1ui,2)
=ygui,2−gui,1ui,2
>)n(σ2
Z+ζ)gui,1ui,2
)n(σ2
Z+ζ)n.(195)
This implies
ygui,1≥)n(σ2
Z+η),(196)
for η<ζ
2.Thatis,ygui,2≥+n(σ2
Z+ζ)implies
ygui,1≥+n(σ2
Z+η).
We deduce that the second integral in the right hand side
of (193) is bounded by
$
y:ygui,1>n(σ2
Z+η)
fZ(ygui1)dy
=$
z:z>n(σ2
Z+η)
fZ(z)dz
=Pr(Z22
Z>nη)
3σ4
Z
2
κ, (197)
for large n, where the first inequality is due to Chebyshev’s
inequality, followed by the substitution of zygui1.
Thus, by (193),
$Di1,g n
t=1
fZ(ytgui1,t)dy
$y∈Di1,g:
ygui,2≤n(σ2
Z+ζ)fZ(ygui1)dy+κ. (198)
Now, we can focus on yRnsuch that
ygui,2≤)n(σ2
Z+ζ).(199)
Then,
fZ(ygui1)fZ(ygui2)
=fZ(ygui1)%1e1
2σ2
Z
ygui22ygui12
&.
(200)
By the triangle inequality,
ygui1≤ygui2+gui1ui2.(201)
Taking the square of both sides, we have
ygui12
≤ygui22+g2ui2ui12
+2ygui2·gui2ui1
≤ygui22+g2α2
n+2n)n(σ2
Z+ζ)
=ygui22+g2α2
n+2g+A(σ2
Z+ζ)
nb,(202)
where the second inequality follows from (191) and (199), and
the equality is due to (192). Thus, for sufficiently large n,
ygui12−ygui22θ. (203)
Hence,
fZ(ygui1)fZ(ygui2)
fZ(ygui1)1e1
2σ2
Z
θ
κfZ(ygui1),(204)
for sufficiently small θ>0, such that 1e1
2σ2
Z
θκ.Now
by (198), we get,
λ1+λ2
Pe,1(i1)+Pe,2(i2,i
1)
=sup
g∈G '1$Di1,g n
t=1
fZ(ytgui1,t)dy(
+sup
g∈G '$Di1,g n
t=1
fZ(ytgui2,t)dy(
sup
g∈G 21$Di1,g n
t=1
fZ(ytgui1,t)dy
+$Di1,g n
t=1
fZ(ytgui2,t)dy3
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
22 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 68, NO. 1, JANUARY 2022
sup
g∈G %1$y∈Di1,g:
ygui,2≤n(σ2
Z+ζ)fZ(ygui1)dyκ
+$Di1,g
fZ(ygui2)dy&
sup
g∈G %1κ$y∈Di1,g:
ygui,2≤n(σ2
Z+ζ)fZ(ygui1)dy
+$y∈Di1,g:
ygui,2≤n(σ2
Z+ζ)fZ(ygui2)dy&
=sup
g∈G %1κ$y∈Di1,g:
ygui,2≤n(σ2
Z+ζ)(fZ(ygui1)
fZ(ygui2))dy&.(205)
Hence, by (204),
λ1+λ2
1κ
κ·inf
g∈G
$y∈Di1,g:
ygui,2≤n(σ2
Z+ζ)fZ(ygui1)dy
12κ, (206)
which leads to a contradiction for sufficiently small κ,such
that 2κ<1λ1λ2. This completes the proof of
Lemma 13.
By Lemma 13 we can define an arrangement of non-
overlapping spheres Sui(n, n)of radius ncentered
at the codewords ui. Since the codewords all belong to a
sphere S0(n, nA)of radius nA centered at the origin,
it follows that the number of packed spheres, i.e., the number
of codewords 2nlog(n)R, is bounded by
2nlog(n)RVo l (S0(n, nA +n))
Vo l (Sui(n, n))
=A+εn
εnn
.(207)
Thus,
R1
log nlog A+εn
εn
=1+b+log 1+ 1
n1+b
log n,(208)
by the same arguments as in the proof for the Gaussian channel
with fast fading (see (148)), which tends to 1+bas n→∞.
Now, since b>0is arbitrarily small, an achievable rate must
satisfy R1. This completes the proof of Theorem 11.
ACKNOWLEDGMENT
The authors gratefully thank Andreas Winter (Universitat
Autònoma de Barcelona) for useful discussions concerning
the scaling of the DI capacity. They also thank Ning Cai
(ShanghaiTech University) for a discussion on the DI capacity
for DMCs. Finally, they also thank Robert Schober (Friedrich
Alexander University) for discussions and questions about the
application of identification theory in molecular communica-
tions, and the anonymous reviewers for helpful comments.
Holger Boche thanks Marios Kountouris for discussions on the
talk and Kai Börner and Dimitar Kroushkov for applications
of deterministic identification.
REFERENCES
[1] M. J. Salariseddigh, U. Pereg, H. Boche, and C. Deppe, “Deterministic
identification over fading channels,” in Proc. IEEE Inf. Theory Workshop
(ITW), Apr. 2021, pp. 1–5.
[2] M. J. Salariseddigh, U. Pereg, H. Boche, and C. Deppe, “Deterministic
identification over channels with power constraints,” in Proc. IEEE Int.
Conf. Commun. (ICC), Jun. 2021, pp. 1–6.
[3] C. E. Shannon, “A mathematical theory of communication,” Bell Syst.
Tec h . J., vol. 27, no. 3, pp. 379–423, Jul./Oct. 1948.
[4] R. Ahlswede and G. Dueck, “Identification via channels,” IEEE Trans.
Inf. Theory, vol. 35, no. 1, pp. 15–29, Jan. 1989.
[5]T.S.Han,Information-Spectrum Methods in Information
Theory (Stochastic Modelling and Applied Probability). Berlin,
Germany: Springer-Verlag, 2014.
[6] S. Derebeyo˘glu, C. Deppe, and R. Ferrara, “Performance analysis of
identification codes,” Entropy, vol. 22, no. 10, p. 1067, Sep. 2020.
[7] G. Fettweis et al., “The tactile internet-ITU-T technology watch report,
Int. Telecom. Union, Geneva, Switzerland, Tech. Rep., 2014.
[8] K. Guan et al., “On the influence of scattering from traffic signs in
vehicle-to-x communications,” IEEE Trans. Veh. Technol., vol. 65, no. 8,
pp. 5835–5849, Aug. 2016.
[9] J. Choi, V. Va, N. G.-Prelcic, R. Daniels, C. R. Bhat, and
R. W. Heath, Jr., “Millimeter-wave vehicular communication to support
massive automotive sensing, IEEE Commun. Mag., vol. 54, no. 12,
pp. 160–167, Dec. 2016.
[10] P. Moulin, “The role of information theory in watermarking and its
application to image watermarking,” Signal Process., vol. 81, no. 6,
pp. 1121–1139, 2001.
[11] Y. Steinberg and N. Merhav, “Identification in the presence of side
information with application to watermarking,” IEEE Trans. Inf. Theory,
vol. 47, no. 4, pp. 1410–1422, May 2001.
[12] R. Ahlswede and N. Cai, Watermarking Identification Codes
With Related Topics on Common Randomness. Berlin,
Germany: Springer-Verlag, 2006, pp. 107–153.
[13] R. Gay, A. Charlesworth, and R. Esen, Online Marketing: A Customer-
Led Approach. London, U.K.: Oxford Univ. Press, 2007.
[14] C. Gur ˘au, “Integrated online marketing communication: Implementation
and management,” J. Commun. Manage., vol. 12, no. 2, pp. 169–184,
May 2008.
[15] Y. Lu, “Industry 4.0: A survey on technologies, applications and open
research issues,” J. Ind. Inf. Integr., vol. 6, pp. 1–10, Jun. 2017.
[16] R. Kaylor, D. Everhart, and J. Lindsay, “Healthcare networks with
biosensors,” U.S. Patent 10 277 170, Apr. 22, 2004.
[17] W. Haselmayr et al., “Integration of molecular communications into
future generation wireless networks, in Proc. 1st 6G Wireless Summit,
Levi, Finland, 2019, pp. 1–2.
[18] J. Cabrera, H. Boche, C. Deppe, R. F. Schaefer, C. Scheunert, and
F. H. Fitzek, “6G and the Post-Shannon theory,” in Shaping Future 6G
Networks: Needs, Impacts and Technologies,N.O.Frederiksenand
H. Gulliksen, Eds. Hoboken, NJ, USA: Wiley, 2021.
[19] R. H. Muller and C. M. Keck, “Challenges and solutions for the
delivery of biotech drugs—A review of drug nanocrystal technology
and lipid nanoparticles,” J. Biotechnol., vol. 113, nos. 1–3, pp. 151–170,
Sep. 2004.
[20] T. Nakano, A. W. Eckford, and T. Haraguchi, Molecular Communica-
tion. Cambridge, U.K.: Cambridge Univ. Press, 2013.
[21] S. K. Hobbs et al., “Regulation of transport pathways in tumor vessels:
Role of tumor type and microenvironment,” Proc. Nat. Acad. Sci. USA,
vol. 95, no. 8, pp. 4607–4612, Apr. 1998.
[22] R. K. Jain, “Transport of molecules, particles, and cells in solid tumors,”
Annu. Rev. Biomed. Eng., vol. 1, no. 1, pp. 241–263, Aug. 1999.
[23] S. Wilhelm et al., “Analysis of nanoparticle delivery to tumours,” Nature
Rev. Mater., vol. 1, no. 5, pp. 1–12, 2016.
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
SALARISEDDIGH et al.: DETERMINISTIC IDENTIFICATION OVER CHANNELS WITH POWER CONSTRAINTS 23
[24] T. Nakano, T. Suda, Y. Okaie, M. J. Moore, and A. V. Vasilakos,
“Molecular communication among biological nanomachines: A layered
architecture and research issues,” IEEE Trans. Nanobiosci., vol. 13,
no. 3, pp. 169–197, Sep. 2014.
[25] H. Boche and C. Deppe, “Robust and secure identification,” in Proc.
IEEE Int. Symp. Inf. Theory (ISIT), Jun. 2017, pp. 1539–1543.
[26] G. Calhoun and G. M. Calhoun, Third Generation Wireless Systems:
Post-Shannon Signal Architectures, vol. 1. Norwood, MA, USA: Artech
House, 2003.
[27] R. Ahlswede, “Elimination of correlation in random codes for arbitrar-
ily varying channels,” Zeitschrift Wahrscheinlichkeitstheorie Verwandte
Gebiete, vol. 44, no. 2, pp. 159–175, 1978.
[28] R. Ahlswede and G. Dueck, “Identification in the presence of feedback-
a discovery of new capacity formulas,” IEEE Trans. Inf. Theory, vol. 35,
no. 1, pp. 30–36, Jan. 1989.
[29] H. Boche, R. F. Schaefer, and H. V. Poor, “On the computability of
the secret key capacity under rate constraints,” in Proc. IEEE Int. Conf.
Acoust. Speech Signal Process. (ICASSP), May 2019, pp. 2427–2431 .
[30] M. V. Burnashev, “On identification capacity of infinite alphabets or
continuous-time channels,” IEEE Trans. Inf. Theory, vol. 46, no. 7,
pp. 2407–2414, Nov. 2000.
[31] H. Boche and C. Deppe, “Secure identification for wiretap channels;
robustness, super-additivity and continuity,” IEEE Trans. Inf. Forensics
Security, vol. 13, no. 7, pp. 1641–1655, Jul. 2018.
[32] A. Winter, “Quantum and classical message identification via quantum
channels,” Quantum Inf. Comput., vol. 5, no. 7, pp. 605–606, 2005.
[33] A. Bracher and A. Lapidoth, “Identification via the broadcast channel,”
IEEE Trans. Inf. Theory, vol. 63, no. 6, pp. 3480–3501, Jun. 2017.
[34] C. E. Shannon, “The zero error capacity of a noisy channel,” IRE Trans.
Inf. Theory, vol. 2, no. 3, pp. 8–19, Sep. 1956.
[35] S. Verdú and V. K. Wei, “Explicit construction of optimal constant-
weight codes for identification via channels,” IEEE Trans. Inf. Theory,
vol. 39, no. 1, pp. 30–36, Jan. 1993.
[36] K. Kurosawa and T. Yoshida, “Strongly universal hashing and identi-
fication codes via channels,” IEEE Trans. Inf. Theory, vol. 45, no. 6,
pp. 2091–2095, Sep. 1999.
[37] J. Bringer, H. Chabanne, G. Cohen, and B. Kindarji, “Private interro-
gation of devices via identification codes,” in Proc. Int. Conf. Cryptol.
India. Springer, 2009, pp. 272–289.
[38] J. Bringer, H. Chabanne, G. Cohen, and B. Kindarji, “Identification
codes in cryptographic protocols,” in Proc. IEEE Inf. Theory Workshop,
Aug. 2010, pp. 1–5.
[39] W. Labidi, “Secure identification for Gaussian channels,” M.S. the-
sis, Dept. Elect. Comput. Eng., LNT, Tech. Univ. Munich, Munich,
Germany, Jun. 2019.
[40] W. Labidi, C. Deppe, and H. Boche, “Secure identification for Gaussian
channels,” in Proc. IEEE Int. Conf. Acoust., Speech Signal Process.
(ICASSP), May 2020, pp. 2872–2876.
[41] W. Labidi, H. Boche, C. Deppe, and M. Wiese, “Identification over the
Gaussian channel in the presence of feedback,” in Proc. IEEE Int. Symp.
Inf. Theory (ISIT), Jul. 2021, pp. 278–283.
[42] R. Ezzine, W. Labidi, H. Boche, and C. Deppe, “Common randomness
generation and identification over Gaussian channels,” in Proc. IEEE
Global Commun. Conf. (GLOBECOM), Dec. 2020, pp. 1–6.
[43] R. Ahlswede and N. Cai, “Identification without randomization,” IEEE
Trans. Inf. Theory, vol. 45, no. 7, pp. 2636–2642, Nov. 1999.
[44] J. J. Ja, “Identification is easier than decoding,” in Proc. 26th Annu.
Symp. Found. Comput. Sci. (SFCS), 1985, pp. 43–50.
[45] M. V. Burnashev, “On the method of types and approximation of output
measures for channels with finite alphabets,” Problemy Inf. Trans.,
vol. 36, no. 3, pp. 195–212, 2000.
[46] R. L. Bocchino, Jr., V. S. Adve, S. V. Adve, and M. Snir, “Parallel
programming must be deterministic by default, in Proc. 1st USENIC
Conf., Hot Topics Parallelism, 2009, p. 4.
[47] E. Arıkan, “Channel polarization: A method for constructing capacity-
achieving codes for symmetric binary-input memoryless channels,” IEEE
Trans. Inf. Theory, vol. 55, no. 7, pp. 3051–3073, Jul. 2009.
[48] R. Ahlswede, “A method of coding and its application to arbitrarily vary-
ing channels,” J. Combinatorics Inf. Sys. Sci., vol. 5, no. 1, pp. 10–35,
1980.
[49] N. Cai, “Personal communication: Professor Cai noted that a straight-
forward application of the proof techniques from A80 does not yield
the DI capacity result,” School Inf. Sci. Technol., ShanghaiTech Univ.,
Shanghai, China, Tech. Rep., Jul. 2020.
[50] J. Stankovic, “Wireless sensor networks,” Computer, vol. 41, no. 10,
pp. 92–95, Oct. 2008.
[51] X. Tian et al., “Wireless body sensor networks based on metamaterial
textiles,” Nature Electron., vol. 2, no. 6, pp. 243–251, Jun. 2019.
[52] H. S. Dhillon, H. Huang, and H. Viswanathan, “Wide-area wireless
communication challenges for the Internet of Things,” IEEE Commun.
Mag., vol. 55, no. 2, pp. 168–174, Feb. 2017.
[53] M. Li, H. Yin, Y. Huang, and Y. Wang, “Impact of correlated fading
channels on cognitive relay networks with generalized relay selection,”
IEEE Access, vol. 6, pp. 6040–6047, 2018.
[54] A. Goldsmith, Wireless Communications. Cambridge, U.K.: Cambridge
Univ. Press, 2005.
[55] L. H. Ozarow, S. Shamai (Shitz), and A. D. Wyner, “Information
theoretic considerations for cellular mobile radio,” IEEE Trans. Veh.
Technol., vol. 43, no. 2, pp. 359–378, May 1994.
[56] E. Biglieri, J. Proakis, and S. Shamai (Shitz), “Fading channels:
Information-theoretic and communications aspects, IEEE Trans. Inf.
Theory, vol. 44, no. 6, pp. 2619–2692, Oct. 1998.
[57] R. Zhang, S. Cui, and Y.-C. Liang, “On ergodic sum capacity of fading
cognitive multiple-access and broadcast channels,” IEEE Trans. Inf.
Theory, vol. 55, no. 11, pp. 5161–5178, Nov. 2009.
[58] R. Masmoudi, “Ergodic capacity for fading channels in cognitive radio
networks,” in Proc. 4th Int. Conf. Adv. Technol. Signal Image Process.
(ATSIP), Mar. 2018, pp. 1–6.
[59] X. Yang, “Capacity of fading channels without channel side informa-
tion,” 2019, arXiv:1903.12360.
[60] A. Nafkha and N. Demni, “Closed-form expressions of ergodic capacity
and MMSE achievable sum rate for MIMO Jacobi and Rayleigh fading
channels,” IEEE Access, vol. 8, pp. 149476–149486, 2020.
[61] M. M. Amiri and D. Gündüz, “Federated learning over wireless
fading channels,” IEEE Trans. Wireless Commun., vol. 19, no. 5,
pp. 3546–3557, May 2020.
[62] D. Tse and P. Viswanath, Fundamentals of Wireless Communication.
Cambridge, U.K.: Cambridge Univ. Press, 2005.
[63] Z. Li, R. Yates, and W. Trappe, Achieving secret communication for
fast Rayleigh fading channels,” IEEE Trans. Wireless Commun.,vol.9,
no. 9, pp. 2792–2799, Sep. 2010.
[64] F. Hosseinigoki and O. Kosut, “Capacity of Gaussian arbitrarily-varying
fading channels,” in Proc. 53rd Annu. Conf. Inf. Sci. Syst. (CISS),
Mar. 2019, pp. 1–6.
[65] K. Besser and E. A. Jorswieck, “Bounds on the ergodic secret-key capac-
ity for dependent fading channels,” in Proc. 24th Int. ITG Workshop
Smart Antennas (WSA), 2020, pp. 1–5.
[66] K. Niu and Y. Li, “Polar codes for fast fading channel: Design
based on polar spectrum,” IEEE Trans. Veh. Technol., vol. 69, no. 9,
pp. 10103–10114, Sep. 2020.
[67] A. J. Goldsmith and P. P. Varaiya, “Capacity of fading channels with
channel side information,” IEEE Trans. Inf. Theory, vol. 43, no. 6,
pp. 1986–1992, Nov. 1997.
[68] V. K. N. Lau and T. Wu, “Optimal transmission and limited feedback
design for OFDM/MIMO systems in frequency selective block fading
channels,” IEEE Trans. Wireless Commun., vol. 6, no. 5, pp. 1569–1573,
May 2007.
[69] A. Vahid, M. A. Maddah-Ali, and A. S. Avestimehr, “Capacity results
for binary fading interference channels with delayed CSIT,” IEEE Trans.
Inf. Theory, vol. 60, no. 10, pp. 6093–6130, Oct. 2014.
[70] I. Csiszár and J. Körner, Information Theory: Coding Theorems for
Discrete Memoryless Systems. New York, NY, USA: Academic, 1982.
[71] S. Shamai (Shitz), “A broadcast strategy for the Gaussian slowly fading
channel,” in Proc. IEEE Int. Symp. Inf. Theory, Jun. 1997, p. 150.
[72] S. Shamai (Shitz) and A. Steiner, “A broadcast approach for a single-
user slowly fading MIMO channel,” IEEE Trans. Inf. Theory, vol. 49,
no. 10, pp. 2617–2635, Oct. 2003.
[73] J. Barros and M. R. D. Rodrigues, “Secrecy capacity of wireless
channels,” in Proc. IEEE Int. Symp. Inf. Theory, Jul. 2006, pp. 356–360.
[74] H. Khoshnevis, I. Marsland, H. Jafarkhani, and H. Yanikomeroglu,
“Space–time signal design for multilevel polar coding in slow fad-
ing broadcast channels,” IEEE Trans. Commun., vol. 67, no. 9,
pp. 5940–5952, Sep. 2019.
[75] K. Eswaran. (2005). Identification Via Channels and
Constant-Weight Codes. [Online]. Available: https://people.eecs.
berkeley.edu/ ananth/229BSpr05/Reports/KrishEswaran.pdf
[76] A. Ahlswede, I. Althöfer, C. Deppe, and U. Tamm, Eds., Identification
and Other Probabilistic Models, Rudolf Ahlswede’s Lectures on Infor-
mation Theory 6 (Foundations in Signal Processing, Communications
and Networking), vol. 15, 1st ed. Springer-Verlag, 2020.
[77] M. J. Salariseddigh, U. Pereg, H. Boche, and C. Deppe, “Deter-
ministic identification over channels with power constraints,” 2020,
arXiv:2010.04239.
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
24 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 68, NO. 1, JANUARY 2022
[78] H. Cohn, “Order and disorder in energy minimization,” in Proc. Int.
Congr. Math. (ICM), vols. 2–4. Singapore: World Scientific, 2010,
pp. 2416–2443.
[79] T. H. Cormen, C. E. Leiserson, R. L. Rivest, and C. Stein, Introduction
to Algorithms. Cambridge, MA, USA: MIT Press, 2009.
[80] T. S. Han and S. Verdú, “New results in the theory of identification via
channels,” IEEE Trans. Inf. Theory, vol. 38, no. 1, pp. 14–25, Jan. 1992.
[81] T. Cover and J. Thomas, Elements of Information Theory (Wiley Series
in Telecommunications). New York, NY, USA: Wiley, 1991.
[82] G. Kramer, Topics in Multi-User Information Theory (Foundations and
Trends in Communications and Information Theory). Boston, MA, USA:
Now, 2008.
[83] R. L. Goodstein, “Transfinite ordinals in recursive number theory,”
J. Symbolic Log., vol. 12, no. 4, pp. 123–129, Dec. 1947.
[84] S. Arora and B. Barak, Computational Complexity: A Modern Approach.
Cambridge, U.K.: Cambridge Univ. Press, 2009.
[85] J. H. Conway and N. J. A. Sloane, Sphere Packings, Lattices and Groups,
vol. 290. New York, NY, USA: Springer, 2013.
[86] Q. Zhang and V. Y. F. Tan, “Covert identification over binary-input
discrete memoryless channels,” 2020, arXiv:2007.13333.
[87] B. A. Bash, D. Goeckel, and D. Towsley, “Limits of reliable communi-
cation with low probability of detection on AWGN channels,” IEEE J.
Sel. Areas Commun., vol. 31, no. 9, pp. 1921–1930, Sep. 2013.
[88] F. Dressler and S. Fischer, “Connecting in-body nano communication
with body area networks: Challenges and opportunities of the internet of
nano things,” Nano Commun. Netw., vol. 6, no. 2, pp. 29–38, Jun. 2015.
[89] I. F. Akyildiz, M. Pierobon, S. Balasubramaniam, and Y. Koucheryavy,
“The internet of bio-nano things,” IEEE Commun. Mag., vol. 53, no. 3,
pp. 32–40, Mar. 2015.
[90] M. V. Burnashev and S. Verdú, “Measures separated in l1metrics and
ID-codes,” Problems Inf. Trans., vol. 30, no. 3, pp. 3–14, 1994.
Mohammad J. Salariseddigh received the B.Sc. degree in electrical engineer-
ing from the K. N. Toosi University of Technology, Tehran, Iran, in 2014, the
M.Sc. degree in communication and multimedia engineering from Friedrich-
Alexander-Universität Erlangen-Nürnberg, Germany, in 2018. He is currently
pursuing the Ph.D. degree with the Institute for Communications Engineering,
Technische Universität München, Germany. He is also working on a joint
research project NewCom—Post Shannon Communication sponsored by the
Federal Ministry of Education and Research with partners from Technische
Universität München and Technische Universität Berlin. His current research
interests include information theory, and coding and identification theory.
Uzi Pereg (Member, IEEE) was born in Jerusalem, Israel, in 1989. He received
the B.Sc. degree (summa cum laude) in electrical engineering from the
Azrieli College of Engineering, Jerusalem, in 2011, and the M.Sc. and Ph.D.
degrees from the Technion—Israel Institute of Technology, Haifa, Israel,
in 2015 and 2019, respectively. Currently, he is a Post-Doctoral Researcher
with the Institute for Communications Engineering, Technical University of
Munich, Munich, Germany. In 2020, he joined the Theory Group, German
Federal Government (BMBF) Project for the design and analysis of quantum
communication and repeater systems. His research interests are in the areas
of quantum communications, information theory, and coding theory. He was a
recipient of the 2018 Pearl Award for Outstanding Research Work in the field
of communications, the 2018 KLA-Tencor Award for Excellent Conference
Paper, the 2018–2019 Viterbi Fellowship, and the 2020–2021 Israel CHE
Fellowship for Quantum Science and Technology.
Holger Boche (Fellow, IEEE) received the Dr.rer.nat. degree in pure mathe-
matics from the Technische Universität Berlin, Berlin, Germany, in 1998, and
the Dipl.Ing. degree in electrical engineering, the degree in mathematics, and
the Dr.Ing. degree in electrical engineering from the Technische Universität
Dresden, Dresden, Germany, in 1990, 1992, and 1994, respectively. He is
currently with the Institute of Theoretical Information Technology, Technische
Universität München, Munich, Germany; the Munich Center for Quantum
Science and Technology (MCQST), Munich; and the Cyber Security in the
Age of Large-Scale Adversaries—Exzellenzcluster (CASA), Ruhr Universität
Bochum, Bochum, Germany. From 1994 to 1997, he was involved in
postgraduate studies in mathematics with the Friedrich-Schiller Universität
Jena, Jena, Germany. In 1997, he joined the Heinrich-HertzInstitut (HHI)
für Nachrichtentechnik Berlin, Berlin. In 2002, he was a Full Professor
of mobile communication networks with the Institute for Communications
Systems, Technische Universität Berlin. In 2003, he became the Director of
the Fraunhofer German-Sino Laboratory for Mobile Communications, Berlin,
and the Director of HHI in 2004. He was a Visiting Professor with ETH
Zurich, Zürich, Switzerland, in Winter 2004 and 2006, and KTH Stockholm,
Stockholm, Sweden, in Summer 2005. Since 2010, he has been with the
Institute of Theoretical Information Technology and a Full Professor with
the Technische Universität München. He was elected as a member of the
German Academy of Sciences Leopoldina in 2008 and the Berlin Brandenburg
Academy of Sciences and Humanities in 2009. Since 2014, he has been a
member and an Honorary Fellow of the TUM Institute for Advanced Study,
Munich. He is a member of the IEEE Signal Processing Society SPCOM
and the SPTM Technical Committee. He was a recipient of the Research
Award Technische Kommunikation from the Alcatel SEL Foundation in 2003,
the Innovation Award from the Vodafone Foundation in 2006, the Gottfried
Wilhelm Leibniz Prize from the German Research Foundation in 2008, and
the 2007 IEEE Signal Processing Society Best Paper Award, and a co-recipient
of the 2006 IEEE Signal Processing Society Best Paper Award.
Christian Deppe (Member, IEEE) received the Dipl.Math. and Dr.Math.
degrees in mathematics from Universität Bielefeld, Bielefeld, Germany,
in 1996 and 1998, respectively. He was a Research and Teaching Assistant
with the Fakultät für Mathematik, Universität Bielefeld, from 1998 to 2010.
From 2011 to 2013, he was the Project Leader of the project Sicherheit und
Robustheit des Quanten-Repeaters of the Federal Ministry of Education and
Research, Fakultät für Mathematik, Universität Bielefeld. In 2014, he was
supported by the DFG Project at the Institute of Theoretical Information
Technology, Technische Universität München. In 2015, he had a temporary
professorship with the Fakultät für Mathematik und Informatik, Friedrich-
Schiller Universität Jena. He was the Project Leader of the project Abhör-
sichere Kommunikation über Quanten-Repeater of the Federal Ministry of
Education and Research, Fakultät für Mathematik, Universität Bielefeld. Since
2018, he has been with the Department of Communications Engineering,
Technische Universität München.
Authorized licensed use limited to: Technische Universitaet Muenchen. Downloaded on February 12,2023 at 21:43:28 UTC from IEEE Xplore. Restrictions apply.
... It was initially observed that deterministic identification leads to much poorer results in terms of code size scaling in the block length [5], [6] than the unrestricted randomized identification. Indeed, DI over discrete memoryless channels can only lead to linear scaling of the message length, log N ∼ Rn [7] as in Shannon's communication paradigm (albeit with a higher rate R). Despite this poorer performance, interest in deterministic codes has recently been renewed, as they have proven to be easier to implement and simulate [8], to explicitly construct [9], and offer reliable single-block performance [5]; see in particular [10] for identification codes. ...
... Also, surprisingly, certain channels with continuous input alphabets have DI codes governed by a slightly superlinear scaling in the block length: log N ∼ Rn log n. This was first observed for Gaussian channels [7], Gaussian channels with both fast and slow fading [11], [12], and Poisson channels [13], [14]. This behaviour differs from that of randomized identification, where the scaling remains the same as for DMCs [15], [16]. ...
... Example 12 (Cf. [7]). Given small enough E(n) > 0, for all sufficiently large n there exists a DI code with E 1 (n), E 2 (n) ≥ E(n) − 3 n for a DMC W : X → Y with |W (X )| pairwise different output probability distributions (equivalently: N row = |W (X )| distinct rows of the channel stochastic matrix) and rate lower-bounded ...
Preprint
Full-text available
We investigate deterministic identification over arbitrary memoryless channels under the constraint that the error probabilities of first and second kind are exponentially small in the block length n, controlled by reliability exponents E1,E20E_1,E_2 \geq 0. In contrast to the regime of slowly vanishing errors, where the identifiable message length scales as Θ(nlogn)\Theta(n\log n), here we find that for positive exponents linear scaling is restored, now with a rate that is a function of the reliability exponents. We give upper and lower bounds on the ensuing rate-reliability function in terms of (the logarithm of) the packing and covering numbers of the channel output set, which for small error exponents E1,E2>0E_1,E_2>0 can be expanded in leading order as the product of the Minkowski dimension of a certain parametrisation the channel output set and logmin{E1,E2}\log\min\{E_1,E_2\}. These allow us to recover the previously observed slightly superlinear identification rates, and offer a different perspective for understanding them in more traditional information theory terms. We further illustrate our results with a discussion of the case of dimension zero, and extend them to classical-quantum channels and quantum channels with tensor product input restriction.
... In practical scenarios, however, randomized encoding is not always feasible. An alternative is a deterministic identification, which also allows for an increase in the number of messages, though the growth, in this case, is only exponential with respect to the block length [4]. ...
... The paper [4] was the first to provide a detailed analysis of deterministic identification over Gaussian channels with power constraints. Before this, it was not known that in this case, the scaling of messages for deterministic identification differs from that of message transmission over these channels. ...
... This result was highly unexpected. Specifically, the following was shown, where L(n, R) denote the scaling: Theorem 1 ( [4]): The deterministic identification (DI) capacity of the Gaussian channel G with power constraints in the 2 n log n -scale, i.e., for L(n, R) = 2 (n log n)R , is bounded by: ...
Preprint
Deterministic identification offers an efficient solution for scenarios where decoding entire messages is unnecessary. It is commonly used in alarm systems and control systems. A key advantage of this approach is that the capacity for deterministic identification in Gaussian channels with power constraints grows superexponentially, unlike Shannon's transmission capacity. This allows for a significantly higher number of messages to be transmitted using this event-driven method. So far, only upper and lower bounds for deterministic identification capacity have been established. Our work introduces a novel construction: galaxy codes for deterministic identification. Using these codes, we demonstrate an improvement in the achievability bound of 1/4 to 3/8, representing a previously unknown advance that opens new possibilities for efficient communication.
... More interestingly, and directly motivating our present work, in certain channels with continuous input alphabets, it was found that deterministic identification codes are governed by a slightly superexponential scaling in block length. Concretely, via fast and slow fading Gaussian channels [9], [13] and over Poisson channels [14] optimal DI codes grow as N (n, λ 1 , λ 1 ) ∼ 2 Rn log n . We are thus motivated to define the slightly superexponential capacity aṡ C DI (W ) := inf λ1,λ2>0 lim inf n→∞ 1 n log n log N DI (n, λ 1 , λ 2 ). ...
... We are thus motivated to define the slightly superexponential capacity aṡ C DI (W ) := inf λ1,λ2>0 lim inf n→∞ 1 n log n log N DI (n, λ 1 , λ 2 ). (7) Indeed, the superexponential DI capacity has been bounded for fast-and slow-fading Gaussian channels G [9], [13] (with a recently improved upper bound in [15]) and for the Poisson channel P [14], [16] as follows: ...
... lim sup n→∞ 1 n log n log N DI (n, λ 1 , λ 2 ). (9) In the present work, we study deterministic identification over a general memoryless channel, mostly with finite output but an arbitrary input alphabet. After reviewing preliminaries about continuous channels, typicality, distance measures, and dimension theory (Section II), in Section III we formulate a certain Hypothesis Testing Lemma III.1 (generalising a previous insight on the intersection of typical sets [19]) which implies that for the construction of a deterministic identification code, it is sufficient to ensure pairwise reliable distinguishability of the output distributions. ...
Article
Full-text available
Following initial work by JaJa, Ahlswede and Cai, and inspired by a recent renewed surge in interest in deterministic identification (DI) via noisy channels, we consider the problem in its generality for memoryless channels with finite output, but arbitrary input alphabets. Such a channel is essentially given by its output distributions as a subset in the probability simplex. Our main findings are that the maximum length of messages thus identifiable scales superlinearly as Rn log n with the block length n, and that the optimal rate R is bounded in terms of the covering (aka Minkowski, or Kolmogorov, or entropy) dimension d of a certain algebraic transformation of the output set: 1/4 d ≤ R ≤ 1/2 d . Remarkably, both the lower and upper Minkowski dimensions play a role in this result. Along the way, we present a Hypothesis Testing Lemma showing that it is sufficient to ensure pairwise reliable distinguishability of the output distributions to construct a DI code. Although we do not know the exact capacity formula, we can conclude that the DI capacity exhibits superactivation: there exist channels whose capacities individually are zero, but whose product has positive capacity. We also generalise these results to classical-quantum channels with finite-dimensional output quantum system, in particular to quantum channels on finite-dimensional quantum systems under the constraint that the identification code can only use tensor product inputs.
... If only deterministic encoding is allowed, then the number of identities that can be identified over a DMC scales exponentially with the blocklength. Nevertheless, the rate is still more significant than the transmission rate in the exponential scale, as shown in [3,4]. ...
... We consider two cases, namely, deterministic and randomized identification. For a transmission system without sensing, it was shown in [1,3] that the number of messages grows exponentially, i.e., N = 2 nC d ID . Remarkably, Theorem 4 demonstrates that by incorporating sensing, the growth rate of the number of messages becomes double exponential (N = 2 2 nC d ID (D) ); this result is notable, and closely parallels the findings on identification with feedback in [14]. ...
Article
Full-text available
In the identification (ID) scheme proposed by Ahlswede and Dueck, the receiver’s goal is simply to verify whether a specific message of interest was sent. Unlike Shannon’s transmission codes, which aim for message decoding, ID codes for a discrete memoryless channel (DMC) are far more efficient; their size grows doubly exponentially with the blocklength when randomized encoding is used. This indicates that when the receiver’s objective does not require decoding, the ID paradigm is significantly more efficient than traditional Shannon transmission in terms of both energy consumption and hardware complexity. Further benefits of ID schemes can be realized by leveraging additional resources such as feedback. In this work, we address the problem of joint ID and channel state estimation over a DMC with independent and identically distributed (i.i.d.) state sequences. State estimation functions as the sensing mechanism of the model. Specifically, the sender transmits an ID message over the DMC while simultaneously estimating the channel state through strictly causal observations of the channel output. Importantly, the random channel state is unknown to both the sender and the receiver. For this system model, we present a complete characterization of the ID capacity–distortion function.
... Deterministic codes, however, are in general easier to implement, simulate, and to explicitly construct [9]- [11]; so, in spite of the poorer performance, an interest was kept in DI codes due to potential applications. This interest was renewed when it was proven that certain channels with continuous input can have superlinear scaling log N ∼ Rn log n of the code size [12], [13]. That is, better than transmission and DI over DMCs, which can only achieve a linear scaling; but worse than general randomized identification, which can achieve an exponential scaling in the code size logarithm log N . ...
Preprint
Full-text available
In our previous work, we presented the Hypothesis Testing Lemma, a key tool that establishes sufficient conditions for the existence of good deterministic identification (DI) codes for memoryless channels with finite output, but arbitrary input alphabets. In this work, we provide a full quantum analogue of this lemma, which shows that the existence of a DI code in the quantum setting follows from a suitable packing in a modified space of output quantum states. Specifically, we demonstrate that such a code can be constructed using product states derived from this packing. This result enables us to tighten the capacity lower bound for DI over quantum channels beyond the simultaneous decoding approach. In particular, we can now express these bounds solely in terms of the Minkowski dimension of a certain state space, giving us new insights to better understand the nature of the protocol, and the separation between simultaneous and non-simultaneous codes. We extend the discussion with a particular channel example for which we can construct an optimum code.
... Similarly, deterministic ID (DI) over memoryless channels with finite output spaces and arbitrary input alphabets has been studied in [16], [17], demonstrating that the number of identifiable messages exhibits super-exponential growth ∼ 2 n log(n)R . Specifically for channels relevant for MC, DI has been studied for binomial channels in [18], for DTPCs in [19], and for fading channels in [20]. Additionally, randomized ID over DTPCs has been investigated in [21]. ...
Preprint
Molecular communication (MC) enables information transfer via molecules, making it ideal for biomedical applications where traditional methods fall short. In many such scenarios, identifying specific events is more critical than decoding full messages, motivating the use of deterministic identification (DI). This paper investigates DI over discrete-time Poisson channels (DTPCs) with inter-symbol interference (ISI), a realistic setting due to channel memory effects. We improve the known upper bound on DI capacity under power constraints from 32+κ\frac{3}{2} + \kappa to 1+κ2\frac{1 + \kappa}{2}. Additionally, we present the first results on deterministic identification with feedback (DIF) in this context, providing a constructive lower bound. These findings enhance the theoretical understanding of MC and support more efficient, feedback-driven biomedical systems.
Chapter
Motivated by deterministic identification via classical channels, where the encoder is not allowed to use randomization, we revisit the problem of identification via quantum channels but now with the additional restriction that the message encoding must use pure quantum states, rather than general mixed states. Together with the previously considered distinction between simultaneous and general decoders, this suggests a two-dimensional spectrum of different identification capacities, whose behaviour could a priori be very different. We demonstrate two new results as our main findings: first, we show that all four combinations (pure/mixed encoder, simultaneous/general decoder) have a double-exponentially growing code size, and that indeed the corresponding identification capacities are lower bounded by the classical transmission capacity for a general quantum channel, which is given by the Holevo-Schumacher-Westmoreland Theorem. Secondly, we show that the simultaneous identification capacity of a quantum channel equals the simultaneous identification capacity with pure state encodings, thus leaving three linearly ordered identification capacities. By considering some simple examples, we finally show that these three are all different: general identification capacity can be larger than pure-state-encoded identification capacity, which in turn can be larger than pure-state-encoded simultaneous identification capacity.
Conference Paper
Full-text available
Due to an increasing number of devices and applications which rely on wireless data transmission, more sensitive data will be processed and transmitted in the future. This makes confidential and secure data communications more important than ever. One way to achieve this goal is physical layer security where the physical properties of the wireless channel are exploited to provide security during the transmission. One new security primitive is secret-key generation. However, the uncertainty with respect to the eavesdropper’s channel and position is one main challenge which heavily influences the performance. In this paper, we derive upper and lower bounds on the ergodic secret-key capacity for fading channels with given marginals and a possible dependence structure. We apply results from mass transportation theory to obtain general bounds and evaluate them for specific fading distributions. In particular, explicit results for Rayleigh and alpha-mu fading channels are presented, including multiple-antenna fading channels.
Article
Full-text available
In this paper, we analyze the construction of identification codes. Identification codes are based on the question: “Is the message I have just received the one I am interested in?”, as opposed to Shannon’s transmission, where the receiver is interested in not only one, but any, message. The advantage of identification is that it allows rates growing double exponentially in the blocklength at the cost of not being able to decode every message, which might be beneficial in certain applications. We focus on a special identification code construction based on two concatenated Reed-Solomon codes and have a closer look at its implementation, analyzing the trade-offs of identification with respect to transmission and the trade-offs introduced by the computational cost of identification codes.
Conference Paper
In this paper, we discuss the potential of integrating molecular communication (MC) systems into future generations of wireless networks. First, we explain the advantages of MC compared to conventional wireless communication using electromagnetic waves at different scales, namely at micro-and macroscale. Then, we identify the main challenges when integrating MC into future generation wireless networks. We highlight that two of the greatest challenges are the interface between the chemical and the cyber (Internet) domain, and ensuring communication security. Finally, we present some future applications, such as smart infrastructure and health monitoring, give a timeline for their realization, and point out some areas of research towards the integration of MC into 6G and beyond
Chapter
Since the breakthrough of Shannon's seminal paper, researchers and engineers have worked on codes and techniques that approach the fundamental limits of message transmission. Given the capacity C of a channel and the block length n of the codewords, the maximum number of possible messages that can be transmitted is 2 nC . In this work, we advocate a paradigm change towards Post-Shannon communication that allows the encoding of up to 2 2 nC messages: a double exponential behavior! This paradigm shift is the study of the transmission of the Gestalt information instead of message-only transmission and involves a shift from the question of what message the sender has transmitted to whether it has transmitted at all, and with the purpose of achieving which goal. Entire careers were built designing methods and codes on top of previous works, bringing only marginal gains in approaching the fundamental limit of Shannon's message transmission. This paradigm change can bring not only marginal but also exponential gains in the efficiency of communication. Within Post-Shannon techniques, we will explore identification codes, the exploitation of resources that are considered useless in the current paradigm such as noiseless feedback common randomness, and the exploitation of multi-channel descriptor information.
Book
The sixth volume of Rudolf Ahlswede's lectures on Information Theory is focused on Identification Theory. In contrast to Shannon's classical coding scheme for the transmission of a message over a noisy channel, in the theory of identification the decoder is not really interested in what the received message is, but only in deciding whether a message, which is of special interest to him, has been sent or not. There are also algorithmic problems where it is not necessary to calculate the solution, but only to check whether a certain given answer is correct. Depending on the problem, this answer might be much easier to give than finding the solution. ``Easier'' in this context means using fewer resources like channel usage, computing time or storage space. Ahlswede and Dueck's main result was that, in contrast to transmission problems, where the possible code sizes grow exponentially fast with block length, the size of identification codes will grow doubly exponentially fast. The theory of identification has now developed into a sophisticated mathematical discipline with many branches and facets, forming part of the Post Shannon theory in which Ahlswede was one of the leading experts. New discoveries in this theory are motivated both by concrete engineering problems and by explorations of the inherent properties of the mathematical structures. Rudolf Ahlswede wrote: It seems that the whole body of present day Information Theory will undergo serious revisions and some dramatic expansions. In this book we will open several directions of future research and start the mathematical description of communication models in great generality. For some specific problems we provide solutions or ideas for their solutions. The lectures presented in this work, which consists of 10 volumes, are suitable for graduate students in Mathematics, and also for those working in Theoretical Computer Science, Physics, and Electrical Engineering with a background in basic Mathematics. The lectures can be used as the basis for courses or to supplement courses in many ways. Ph.D. students will also find research problems, often with conjectures, that offer potential subjects for a thesis. More advanced researchers may find questions which form the basis of entire research programs. The book also contains an afterword by Gunter Dueck.
Article
This paper considers the covert identification problem in which a sender aims to reliably convey an identification (ID) message to a set of receivers via a binary-input discrete memoryless channel (BDMC), and simultaneously to guarantee that the communication is covert with respect to a warden who monitors the communication via another independent BDMC. We prove a square-root law for the covert identification problem. This states that an ID message of size exp(exp(Θ(n)))\exp (\exp (\Theta (\sqrt {\text {n}}))) can be transmitted over n channel uses. We then characterize the exact pre-constant in the Θ()\Theta (\cdot) notation. This constant is referred to as the covert identification capacity. We show that it equals the recently developed covert capacity in the standard covert communication problem, and somewhat surprisingly, the covert identification capacity can be achieved without any shared key between the sender and receivers. The achievability proof relies on a random coding argument with pulse-position modulation (PPM), coupled with a second stage which performs code refinements. The converse proof relies on an expurgation argument as well as results for channel resolvability with stringent input constraints.