Skip to article frontmatterSkip to article content

Channel Capacity

Channel capacity represents the fundamental limit of a communication channel, defining the maximum rate at which information can be transmitted with an arbitrarily small probability of error.

Channel Capacity of DMC

We consider a discrete memoryless channel (DMC), such as the binary symmetric channel (BSC) with crossover probability pp (the probability of a bit being flipped).

The capacity is closely tied to the channel’s error characteristics.

Specifically, for the BSC, the capacity is

C=1Hb(p),C = 1 - H_b(p),

where

Hb(p)=plog2p(1p)log2(1p)H_b(p) = -p \log_2 p - (1 - p) \log_2 (1 - p)

is the binary entropy function.

Recall that Hb(p)H_b(p) measures the uncertainty in a binary random variable with probability pp.

Consequently, 1Hb(p)1 - H_b(p) quantifies the maximum rate (in bits per channel use) at which reliable communication is possible over the BSC, capturing the reduction in uncertainty about the input given the output.

General Definition of Channel Capacity

More generally, the capacity CC of any channel is defined as

C=maxpI(X;Y),\boxed{C = \max_{\vec{p}} I(X; Y),}

where I(X;Y)I(X; Y) is the mutual information between the input random variable XX (with alphabet X\mathcal{X}) and the output random variable YY (with alphabet Y\mathcal{Y}).

Formally,

I(X;Y)=xXyYp(x,y)log2p(x,y)p(x)p(y),I(X; Y) = \sum_{x \in \mathcal{X}} \sum_{y \in \mathcal{Y}} p(x, y) \log_2 \frac{p(x, y)}{p(x) p(y)},

where p(x,y)=p(x)P[yx]p(x,y) = p(x) P[y \mid x], p(x)p(x) is the input probability mass function (PMF), and P[yx]P[y \mid x] is the conditional probability of yy given xx.

The maximization is performed over all possible input PMFs p=(p1,p2,,pX)\vec{p} = (p_1, p_2, \ldots, p_{|\mathcal{X}|}), subject to:

pi0(i=1,2,,X),i=1Xpi=1,p_i \ge 0 \quad (i = 1, 2, \ldots, |\mathcal{X}|), \qquad \sum_{i=1}^{|\mathcal{X}|} p_i = 1,

ensuring p\vec{p} is a valid PMF.

The resulting capacity CC, in bits per channel use (assuming base-2 logarithms), is the highest achievable rate at which the error probability can be made arbitrarily small through appropriate coding.

Units of Channel Capacity

If the channel transmits one symbol every tst_s seconds, then the capacity in bits per second or nats per second is Cts\frac{C}{t_s}.

This expresses the rate on a time-based rather than a per-use basis, which is important for continuous transmission.

Shannon’s Noisy Channel Coding Theorem

Shannon’s Second Theorem (1948), also known as the Noisy Channel Coding Theorem, underpins the concept of channel capacity:

Noisy Channel Coding Theorem

Reliable communication over a DMC is possible if the transmission rate RR (in bits per channel use) is less than CC.
Conversely, if R>CR > C, reliable communication is impossible, because the error probability cannot be made arbitrarily small regardless of the coding scheme.

It means that there exists a coding scheme (often with sufficiently long block codes) that achieves negligible error for R<CR < C, but no such scheme exists for R>CR > C.

This result establishes CC as the ultimate limit for reliable communication and provides a theoretical benchmark against which practical systems are measured.

It also guides the design of error-correcting codes and modulation schemes, ensuring transmission rates do not exceed the channel’s fundamental capabilities.

Example: Capacity of the BSC

This example is based on [Proakis’s book, Example 6.5-1].

For the BSC with X=Y={0,1}\mathcal{X} = \mathcal{Y} = \{0, 1\} and crossover probability pp (the probability of a bit being flipped), the capacity is maximized by using a uniform input distribution, i.e., P[X=0]=P[X=1]=12P[X = 0] = P[X = 1] = \frac{1}{2}.

Due to the channel’s symmetry:

I(X;Y)=H(Y)H(YX).I(X; Y) = H(Y) - H(Y \mid X).

Hence,

C=I(X;Y)=H(Y)H(YX)=1Hb(p).C = I(X; Y) = H(Y) - H(Y \mid X) = 1 - H_b(p).

Expanding Hb(p)H_b(p) as plog2p(1p)log2(1p)-p \log_2 p - (1 - p)\log_2 (1 - p), we get

C=1+plog2p+(1p)log2(1p).C = 1 + p \log_2 p + (1 - p) \log_2 (1 - p).

Special Cases

When plotted against the signal-to-noise ratio (SNR) per bit, typically denoted Eb/N0\mathcal{E}_b/N_0, the capacity CC increases monotonically, indicating better reliability as SNR rises.

In practical analyses (e.g., BPSK modulation), the crossover probability pp is related to SNR via the error function.

Capacity of the Discrete-Time Binary-Input AWGN Channel

We consider a binary-input additive white Gaussian noise (AWGN) channel with inputs X=±AX = \pm A (e.g., BPSK signaling) and continuous output Y=X+NY = X + N, where NN(0,σ2)N \sim \mathcal{N}(0,\sigma^2).

The conditional probability density function (PDF) is

p(yx)=12πσ2exp((yx)22σ2),x=±A.p(y \mid x) = \frac{1}{\sqrt{2\pi \sigma^2}} \exp \Bigl(-\frac{(y - x)^2}{2\sigma^2}\Bigr), \quad x = \pm A.

By symmetry, the capacity is again maximized by a uniform input PMF: P[X=A]=P[X=A]=12P[X = A] = P[X = -A] = \frac{1}{2}.

Then, the channel capacity is

C=I(X;Y)=H(Y)H(YX).C = I(X; Y) = H(Y) - H(Y \mid X).

However, computing CC directly from the continuous output YY is more conveniently done using:

I(X;Y)=x{A,A}P(x)p(yx)log2(p(yx)p(y))dy,I(X; Y) = \sum_{x \in \{A, -A\}} P(x) \int_{-\infty}^{\infty} p(y \mid x) \log_2 \Bigl(\frac{p(y \mid x)}{p(y)}\Bigr) dy,

with P(x)=12P(x) = \frac{1}{2}.

Thus,

C=12p(yA)log2(p(yA)p(y))dy+12p(yA)log2(p(yA)p(y))dy,\begin{split} C &= \frac{1}{2} \int_{-\infty}^{\infty} p(y \mid A) \log_2 \Bigl(\frac{p(y \mid A)}{p(y)}\Bigr) dy \\ &\quad+ \frac{1}{2} \int_{-\infty}^{\infty} p(y \mid -A) \log_2 \Bigl(\frac{p(y \mid -A)}{p(y)}\Bigr) dy, \end{split}

where

p(y)=12p(yA)+12p(yA)p(y) = \frac{1}{2} p(y \mid A) + \frac{1}{2} p(y \mid -A)

is the overall output PDF.

Simplified Form and Behavior

While the integral does not have a simple closed-form solution, it can be written as

C=12g(Aσ)+12g(Aσ),C = \frac{1}{2} g \Bigl(\frac{A}{\sigma}\Bigr) + \frac{1}{2} g \Bigl(-\frac{A}{\sigma}\Bigr),

where

g(x)=12πexp((ux)22)log2(21+exp(2ux))du,g(x) = \int_{-\infty}^{\infty} \frac{1}{\sqrt{2\pi}} \exp \Bigl(-\frac{(u - x)^2}{2}\Bigr) \log_2 \Bigl(\frac{2}{1 + \exp(-2 u x)}\Bigr) du,

and x=Aσx = \frac{A}{\sigma} is the signal-to-noise amplitude ratio.

Due to symmetry (g(x)g(-x) relates closely to g(x)g(x)), the capacity depends on A2σ2\frac{A^2}{\sigma^2}, often expressed as EbN0\frac{\mathcal{E}_b}{N_0} (energy per bit to noise power spectral density ratio) in communication systems.

As EbN0\frac{\mathcal{E}_b}{N_0} increases, CC grows from 0 (at very low SNR) to 1 bit per symbol (at high SNR), reflecting the channel’s ability to approach error-free operation.

For example, for target rates R=12R = \frac{1}{2} or 13\frac{1}{3}, specific EbN0\frac{\mathcal{E}_b}{N_0} thresholds (e.g., 0.188 dB and -0.496 dB) indicate the minimum required SNR—these values are typically found via numerical methods or simulation.

Capacity of Symmetric Channels

In certain channel models, such as the Binary Symmetric Channel (BSC) and the binary-input AWGN channel, the channel transition probabilities exhibit symmetry.

A symmetric channel is characterized by a transition probability matrix P=[P(yjxi)]\mathbf{P} = [P(y_j \mid x_i)] that satisfies:

As an example, the BSC with crossover probability pp has transition matrix

P=[1ppp1p].\mathbf{P} = \begin{bmatrix} 1 - p & p\\ p & 1 - p \end{bmatrix}.

Row 1, (1p,p)(1 - p, p), is a permutation of Row 2, (p,1p)(p, 1 - p), and similarly for the columns.

This symmetry implies that the channel treats all inputs equivalently up to a possible relabeling of outputs.

Capacity of Symmetric Channels Under a Uniform Input

For symmetric channels, the capacity-achieving input distribution is uniform, i.e., P[xi]=1/XP[x_i] = 1 / |\mathcal{X}| for each xiXx_i \in \mathcal{X}.

The resulting capacity is

C=log2YH(p),C = \log_2 \bigl|\mathcal{Y}\bigr| - H(\vec{p}),

where Y\bigl|\mathcal{Y}\bigr| is the size of the output alphabet, and p=(p1,p2,,pY)\vec{p} = (p_1, p_2, \ldots, p_{|\mathcal{Y}|}) is any row of the transition matrix P\mathbf{P}.

Because each row is just a permutation of the others, p\vec{p} has the same entropy regardless of which row is chosen:

H(p)=j=1Ypjlog2pj.H(\vec{p}) = -\sum_{j=1}^{|\mathcal{Y}|} p_j \log_2 p_j.

Thus, for the BSC (X=Y={0,1}\mathcal{X} = \mathcal{Y} = \{0,1\}) with crossover probability pp, each row is (p,1p)(p, 1 - p) or (1p,p)(1 - p, p). Since Y=2\bigl|\mathcal{Y}\bigr| = 2, we get

C=log2(2)Hb(p)=1[plog2p(1p)log2(1p)]=1Hb(p),C = \log_2(2) - H_b(p) = 1 - \bigl[-p \log_2 p - (1 - p) \log_2(1 - p)\bigr] = 1 - H_b(p),

confirming the known BSC capacity.

It is noted that output symmetry (as in binary-input AWGN) is sufficient to guarantee a uniform capacity-achieving input, even if the channel does not have a finite transition matrix.

Conditions for Capacity-Achieving Distributions

For a general discrete memoryless channel (DMC), the capacity

C=maxP[x]I(X;Y)\boxed{ C = \max_{P[x]} I(X; Y) }

is achieved by an input distribution {P[x]}\{P[x]\} that satisfies the following (sometimes referred to as the Kuhn–Tucker conditions in information theory):

Here, the pointwise mutual information is

I(x;Y)=yYP(yx)log2(P(yx)P[y]),I(x; Y) = \sum_{y \in \mathcal{Y}} P(y \mid x) \log_2 \biggl(\frac{P(y \mid x)}{P[y]}\biggr),

and

P[y]=xXP[x]P(yx).P[y] = \sum_{x' \in \mathcal{X}} P[x'] P(y \mid x').

These conditions ensure that:

In symmetric channels, the uniform distribution satisfies these conditions because I(x;Y)I(x; Y) is the same for all xx.

For non-symmetric channels, finding the optimal distribution often requires iterative methods (e.g., the Blahut–Arimoto algorithm) if uniform inputs do not satisfy the conditions.

Capacity of the Discrete-Time AWGN Channel (Power-Constrained)

Consider a discrete-time AWGN channel:

Yi=Xi+Ni,Y_i = X_i + N_i,

where NiN_i are i.i.d. Gaussian random variables with mean 0 and variance σ2\sigma^2.

The inputs {Xi}\{X_i\} must satisfy the average power constraint

E[X2]    P.\mathbb{E}[X^2] \;\le\; P.

Sphere Packing Argument

For nn uses of the channel, the transmitted vector x=(x1,,xn)\vec{x} = (x_1, \ldots, x_n) must lie within an nn-dimensional sphere of radius nP\sqrt{nP}.

Because the noise vector n=(n1,,nn)\vec{n} = (n_1,\ldots,n_n) is Gaussian, the received vector

y=x+n\vec{y} = \vec{x} + \vec{n}

lies in a sphere of radius n(P+σ2)\sqrt{n(P + \sigma^2)}.

Approximating the maximum number of distinguishable codewords (messages) by the ratio of volumes leads to

M=(1+Pσ2)n/2.M = \Bigl(1 + \frac{P}{\sigma^2}\Bigr)^{ n/2}.

Hence, the achievable rate (in bits per channel use) is

R=log2Mn=12log2(1+Pσ2).R = \frac{\log_2 M}{n} = \frac{1}{2} \log_2 \Bigl(1 + \frac{P}{\sigma^2}\Bigr).

This is precisely the channel capacity for an average power constraint PP.

Achievability follows by choosing XX to be Gaussian N(0,P)\mathcal{N}(0, P), which maximizes I(X;Y)I(X;Y).

Formally,

I(X;Y)=h(Y)h(YX)=h(Y)h(N),I(X;Y) = h(Y) - h(Y\mid X) = h(Y) - h(N),

where h()h(\cdot) denotes differential entropy, and XN(0,P)X\sim\mathcal{N}(0,P) maximizes h(Y)h(Y).

It is noted that the capacity is in bits per channel use, and can be converted to bits per second by multiplying with the sampling rate.

Capacity of the Band-Limited Waveform AWGN Channel (Power-Constrained)

Now consider a continuous-time AWGN channel with:

It can be shown (via sampling or dimensionality arguments) that this setup is equivalent to 2W2W discrete-time uses per second of an AWGN channel with:

Per-Use Capacity

The discrete-time channel capacity per use is

Cdiscrete=12log2(1+P2WN02)=12log2(1+PN0W).C_{\mathrm{discrete}} = \frac{1}{2} \log_2 \Bigl(1 + \frac{\frac{P}{2W}}{\frac{N_0}{2}}\Bigr) = \frac{1}{2} \log_2 \Bigl(1 + \frac{P}{N_0 W}\Bigr).

Because there are 2W2W uses per second, the total channel capacity in bits per second is

C=(2W)×Cdiscrete=Wlog2(1+PN0W).C = (2W) \times C_{\mathrm{discrete}} = W \log_2 \Bigl(1 + \frac{P}{N_0 W}\Bigr).

This is Shannon’s well-known capacity formula for a band-limited AWGN channel.

The term PN0W\frac{P}{N_0 W} is the signal-to-noise ratio (SNR).

Capacity Behavior

Bandwidth Efficiency of Band-Limited Waveform AWGN Channel with Power Constraint

The bandwidth efficiency of a communication system quantifies how effectively the available bandwidth is utilized to transmit data.

It is defined as the ratio of the bit rate RR (in bits per second, b/s) to the bandwidth WW (in Hertz, Hz):

r=RW(bits per second per Hertz, b/s/Hz).\boxed{ r = \frac{R}{W} \quad (\text{bits per second per Hertz, b/s/Hz}). }

Often denoted by rr, this metric indicates the number of bits transmitted per unit of bandwidth and thus provides insight into the spectral efficiency of a signaling scheme.

To investigate the trade-off between bandwidth efficiency and power efficiency, we begin with the capacity of a band-limited AWGN channel under power constraint PP and noise power spectral density N0/2N_0/2:

C=Wlog2(1+PN0W)(bits per second).\boxed{ C = W \log_2 \Bigl(1 + \frac{P}{N_0 W}\Bigr) \quad (\text{bits per second}). }

For reliable communication, the bit rate RR must satisfy R<CR < C, ensuring that errors can be made arbitrarily small through suitable coding techniques.

Substituting the expression for CC into R<CR < C:

R<Wlog2(1+PN0W).R < W \log_2 \Bigl(1 + \frac{P}{N_0 W}\Bigr).

Dividing both sides by the bandwidth WW yields

RW<log2(1+PN0W).\frac{R}{W} < \log_2 \Bigl(1 + \frac{P}{N_0 W}\Bigr).

Since r=RWr = \frac{R}{W}, we obtain

r<log2(1+PN0W).r < \log_2 \Bigl(1 + \frac{P}{N_0 W}\Bigr).

This inequality reveals a fundamental relationship between the bandwidth efficiency rr and the SNR PN0W\frac{P}{N_0 W}, identifying the maximum achievable rr for given values of power PP and bandwidth WW.

Relating Bandwidth Efficiency to Power Efficiency

To express this connection in terms of power efficiency, we introduce the energy per bit to noise power spectral density ratio, EbN0\frac{\mathcal{E}_b}{N_0}.

For a signaling scheme with bit rate RR and power PP, the total energy per symbol E\mathcal{E} is linked to the symbol duration Ts=1RT_s = \frac{1}{R} (assuming one symbol per bit for simplicity):

E=PTs=PR.\mathcal{E} = P T_s = \frac{P}{R}.

For MM-ary signaling, R=log2MTsR = \frac{\log_2 M}{T_s}, but here we focus directly on the bit rate.

In that case,

EbN0=Elog2M=PR.\frac{\mathcal{E}_b}{N_0} = \frac{\mathcal{E}}{\log_2 M} = \frac{P}{R}.

Hence,

P=REbN0.P = R \frac{\mathcal{E}_b}{N_0}.

Substitute PP into the earlier capacity inequality r<log2(1+PN0W)r < \log_2 \bigl(1 + \frac{P}{N_0 W}\bigr):

r<log2(1+REbN0N0W)=log2(1+rEbN0),r < \log_2 \Bigl(1 + \frac{R \frac{\mathcal{E}_b}{N_0}}{N_0 W}\Bigr) = \log_2 \Bigl(1 + \frac{r \mathcal{E}_b}{N_0}\Bigr),

since RW=r\frac{R}{W} = r. Rearranging this:

1+rEbN0>2rrEbN0>2r1EbN0>2r1r.1 + \frac{r \mathcal{E}_b}{N_0} > 2^r \quad\Rightarrow\quad \frac{r \mathcal{E}_b}{N_0} > 2^r - 1 \quad\Rightarrow\quad \frac{\mathcal{E}_b}{N_0} > \frac{2^r - 1}{r}.

This formula specifies the minimum EbN0\frac{\mathcal{E}_b}{N_0} required for reliable communication at a given bandwidth efficiency rr.

It demonstrates that higher rr (i.e., more bits per Hz) necessitates greater power efficiency, underscoring the trade-off between spectral resources and power resources.

Minimum EbN0\frac{\mathcal{E}_b}{N_0} for Reliable Communication

To identify the absolute minimum EbN0\frac{\mathcal{E}_b}{N_0} that enables reliable transmission, we examine the limiting case r0r \to 0 (vanishingly small bandwidth efficiency). From

EbN0>2r1r,\frac{\mathcal{E}_b}{N_0} > \frac{2^r - 1}{r},

use the Taylor expansion 2r=erln21+rln22^r = e^{r \ln 2} \approx 1 + r \ln 2 for small rr. Thus,

2r1rln2limr02r1r=limr0rln2r=ln2.2^r - 1 \approx r \ln 2 \quad\Rightarrow\quad \lim_{r \to 0} \frac{2^r - 1}{r} = \lim_{r \to 0} \frac{r \ln 2}{r} = \ln 2.

Therefore,

EbN0>ln2    0.693    1.6dB.\frac{\mathcal{E}_b}{N_0} > \ln 2 \;\approx\; 0.693 \;\approx\; -1.6 \text{dB}.

This is known as the Shannon limit, representing the fundamental boundary below which no communication system can achieve arbitrarily low error probability.

Converting to decibels via 10log10(0.693)1.6dB10 \log_{10}(0.693) \approx -1.6 \text{dB}, we see this limit applies to any physical communication scheme.

Reaching the limit requires r0r \to 0, which corresponds to WW \to \infty when RR is fixed; effectively, one leverages vast bandwidth to reduce the SNR per Hz but compensates with sophisticated coding across many dimensions.

Achieving Capacity With Orthogonal Signals

Next, consider a system employing MM orthogonal waveforms (e.g., frequency-shift keying, FSK) over an AWGN channel.

The probability of error PeP_e when distinguishing among MM signals, each having energy E=PTs\mathcal{E} = P T_s, is

Pe=12π[1(1Q(x))M1]exp((x2E/N0)22)dx,P_e = \frac{1}{\sqrt{2\pi}} \int_{-\infty}^{\infty} \Bigl[1 - \bigl(1 - Q(x)\bigr)^{M-1}\Bigr] \exp \Bigl(-\frac{\bigl(x - \sqrt{2\mathcal{E}/N_0}\bigr)^2}{2}\Bigr) dx,

where

Q(x)=12πxexp(t22)dtQ(x) = \frac{1}{\sqrt{2\pi}} \int_x^\infty \exp \Bigl(-\frac{t^2}{2}\Bigr) dt

is the Gaussian tail function, and 2EN0\sqrt{\frac{2\mathcal{E}}{N_0}} is the signal amplitude that depends on the SNR.

Since this integral typically lacks a closed-form solution, it is evaluated numerically to assess error performance as a function of EN0\frac{\mathcal{E}}{N_0}.

Infinite Bandwidth Scenario

When WW \to \infty, the channel capacity approaches

C=PN0ln2,C_\infty = \frac{P}{N_0 \ln 2},

as derived from the infinite-bandwidth limit.

If we let MM \to \infty (with TT denoting the signaling duration), the bit rate RR can approach various fractions of CC_\infty.

Upper bounds on PeP_e are often given for two regimes:

Pe<{2×2T(12CR),0R14C,2×2T(CR)2,14CRC.P_e < \begin{cases} 2 \times 2^{- T \bigl(\frac{1}{2}C_\infty - R\bigr)}, & 0 \le R \le \frac{1}{4}C_\infty,\\ 2 \times 2^{- T \bigl(\sqrt{C_\infty} - \sqrt{R}\bigr)^2}, & \frac{1}{4}C_\infty \le R \le C_\infty. \end{cases}

Since C=PN0ln2C_\infty = \frac{P}{N_0 \ln 2}, reliable communication is possible whenever R<CR < C_\infty, and the error probability PeP_e decreases exponentially as TT increases.

Orthogonal signaling can achieve this capacity in the limit MM \to \infty by exploiting the large-dimensional signal space, thereby approaching the Shannon limit EbN0=ln2\frac{\mathcal{E}_b}{N_0} = \ln 2.

The Channel Reliability Function

In digital communication over an infinite-bandwidth additive white Gaussian noise (AWGN) channel using (M)(M)-ary orthogonal signals (e.g., frequency-shift keying), the probability of error PeP_e decreases exponentially with the signaling duration (T)(T).

This behavior is captured by exponential bounds that provide insight into how reliably information can be transmitted at a given rate.

For such a system, an upper bound on the error probability is expressed as:

Pe<2×2TE(R).P_e < 2 \times 2^{-T E(R)}.

Here, (T)(T) is the duration of the signal (related to (M)(M) and the bit rate (R)(R) via M=2RTM = 2^{R T} for binary encoding), and E(R)E(R) is the channel reliability function that dictates the exponential rate of decay of PeP_e as (T)(T) increases.

The reliability function depends on the communication rate (R)(R) (in bits per second) and the channel’s infinite-bandwidth capacity

C=PN0ln2C_{\infty} = \frac{P}{N_0 \ln 2}

(derived previously in the limit WW \to \infty), where PP is the signal power and N0/2N_0/2 is the noise power spectral density.

Specifically,

E(R)={12CR,0R14C,(CR)2,14CRC.E(R) = \begin{cases} \frac{1}{2} C_{\infty} - R, & \quad 0 \leq R \leq \frac{1}{4} C_{\infty}, \\ \bigl(\sqrt{C_{\infty}} - \sqrt{R}\bigr)^2, & \quad \frac{1}{4} C_{\infty} \leq R \leq C_{\infty}. \end{cases}

Low-Rate Regime (0R14C)\bigl(0 \leq R \leq \frac{1}{4} C_{\infty}\bigr)

The exponent is linear, E(R)=12CRE(R) = \frac{1}{2} C_{\infty} - R, indicating that PeP_e decays exponentially as long as R<12CR < \frac{1}{2} C_{\infty}.

At R=14CR = \frac{1}{4} C_{\infty}, E(R)=14CE(R) = \frac{1}{4} C_{\infty}.

High-Rate Regime (14CRC)\bigl(\frac{1}{4} C_{\infty} \leq R \leq C_{\infty}\bigr)

The exponent becomes quadratic,

E(R)=(CR)2,E(R) = \bigl(\sqrt{C_{\infty}} - \sqrt{R}\bigr)^2,

reflecting a slower decay as RR approaches CC_{\infty}.

At R=CR = C_{\infty}, E(R)=0E(R) = 0, and PeP_e no longer decreases exponentially.

Comparison With the Union Bound

For comparison, the union bound on PeP_e provides a simpler but looser exponential factor [Proakis, Eq. 4.4-17]:

Pe12×2T(12CR),0R12C.P_e \leq \frac{1}{2} \times 2^{-T \bigl(\frac{1}{2} C_{\infty} - R\bigr)}, \quad 0 \leq R \leq \frac{1}{2} C_{\infty}.

This bound uses a linear exponent, 12CR\frac{1}{2} C_{\infty} - R, which is valid up to R=12CR = \frac{1}{2} C_{\infty}, where it becomes zero.

Beyond this point, the union bound becomes trivial (Pe1P_e \leq 1), lacking the nuanced behavior of E(R)E(R) in the higher-rate regime.

The looseness arises because the union bound overestimates error events by summing pairwise error probabilities, whereas E(R)E(R) more accurately accounts for their joint probability.

Exponential Tightness of Gallager’s Bound

The bound, derived by Gallager (1965), is exponentially tight, meaning it represents the best possible exponential upper bound.

There is no alternative reliability function E1(R)\mathcal{E}_1(R) with E1(R)>E(R)\mathcal{E}_1(R) > E(R) for any RR that still bounds PeP_e correctly.

This tightness establishes E(R)E(R) as the definitive measure of error decay for the infinite-bandwidth AWGN channel.

Matching Upper and Lower Bounds

The error probability is bounded both above and below:

Kl2TE(R)PeKu2TE(R),\boxed{ K_l 2^{-T E(R)} \leq P_e \leq K_u 2^{-T E(R)}, }

where Ku=2K_u = 2 is the upper bound constant, and KlK_l is a positive constant ensuring that E(R)E(R) applies to both bounds.

These constants KlK_l and KuK_u have only a weak dependence on TT, as shown by:

limT1TlnKl=limT1TlnKu=0.\boxed{ \lim_{T \to \infty} \frac{1}{T} \ln K_l = \lim_{T \to \infty} \frac{1}{T} \ln K_u = 0. }

Hence, lnKl\ln K_l and lnKu\ln K_u grow sublinearly with TT (e.g., as constants or lnT\ln T), so their effect on the exponent becomes negligible for large TT.

The dominant term in the exponent is TE(R)-T E(R), confirming that E(R)E(R) governs the asymptotic error behavior.

Because orthogonal signals are asymptotically optimal as MM \to \infty (approaching the capacity CC_{\infty}), this reliability function applies universally.

The lower bound implies that no signaling scheme can achieve a faster exponential decay than E(R)E(R), making it a fundamental characteristic of digital signaling over the infinite-bandwidth AWGN channel.

For R<CR < C_{\infty}, E(R)>0E(R) > 0, and therefore Pe0P_e \to 0 as TT \to \infty, in alignment with the noisy channel coding theorem.