Skip to article frontmatterSkip to article content

Channel Modeling

End-to-End System Overview

Channel Encoding

In digital communication systems, the discrete channel encoder plays a critical role in preparing a binary information sequence for transmission over a noisy channel.

The primary function of the encoder is to introduce redundancy into the binary sequence in a controlled and systematic manner.

This redundancy refers to additional bits that do not carry new information but are strategically added to enable the receiver to detect and correct errors caused by noise and interference during transmission.

Without such redundancy, the receiver would have no means to distinguish between the intended signal and the distortions introduced by the channel, rendering reliable communication impossible in the presence of noise.

The encoding process involves segmenting the input binary information sequence into blocks of kk bits, where kk represents the number of information bits per block.

Each unique kk-bit sequence is then mapped to a corresponding nn-bit sequence, known as a codeword, where

n>k\boxed{n > k}

This mapping ensures that each possible kk-bit input has a distinct nn-bit output, preserving the uniqueness of the information while embedding redundancy.

For example, a simple repetition code might map a single bit (e.g., k=1k = 1, input (0)(0)) to a three-bit codeword (e.g., n=3n = 3, output (000)(000)), repeating the bit to add redundancy.

Code Rate

The amount of redundancy introduced by this encoding process is quantified by the ratio n/kn/k.

This ratio indicates how many total bits (nn) are transmitted for each information bit (kk).

A higher n/kn/k implies more redundancy; for instance, if k=4k = 4 and n=7n = 7, then

nk=74=1.75\frac{n}{k} = \frac{7}{4} = 1.75

meaning 1.75 bits are sent per information bit, with 0.75 bits being redundant.

The reciprocal of this ratio, k/nk/n, is defined as the code rate, denoted RcR_c:

Rc=kn\boxed{R_c = \frac{k}{n}}

The code rate measures the efficiency of the encoding scheme, representing the fraction of the transmitted bits that carry actual information.

For the example above (k=4k = 4, n=7n = 7),

Rc=470.571R_c = \frac{4}{7} \approx 0.571

meaning approximately 57.1% of the transmitted bits are information, and the remaining 42.9% are redundancy.

A code rate closer to 1 indicates less redundancy (higher efficiency), while a lower RcR_c indicates more redundancy (better error protection but lower efficiency).

The choice of RcR_c balances the trade-off between data throughput and error resilience, depending on the channel’s noise characteristics.

Modulation and Interface to the Channel

The binary sequence output from the channel encoder, consisting of nn-bit codewords, is passed to the modulator, which serves as the interface between the digital system and the physical communication channel (e.g., a wireless medium or optical fiber).

The modulator’s role is to convert the discrete binary sequence into a continuous-time waveform suitable for transmission over the channel.

In its simplest form, the modulator employs binary modulation, where each bit in the sequence is mapped to one of two distinct waveforms:

For example, in binary phase-shift keying (BPSK), s1(t)s_1(t) and s2(t)s_2(t) might be sinusoidal signals differing in phase (e.g., s1(t)=Acos(2πfct)s_1(t) = A \cos(2\pi f_c t) and s2(t)=Acos(2πfct)s_2(t) = -A \cos(2\pi f_c t)), transmitted over a symbol duration TT.

This one-to-one mapping occurs at a rate determined by the bit rate of the encoded sequence.

Alternatively, the modulator can operate on blocks of qq bits at a time, using MM-ary modulation, where M=2qM = 2^q represents the number of possible waveforms.

Each unique qq-bit block is mapped to one of MM distinct waveforms.

For instance, if q=2q = 2, then M=22=4M = 2^2 = 4, and the modulator might use four waveforms (e.g., in quadrature phase-shift keying, QPSK), such as s1(t),s2(t),s3(t),s4(t)s_1(t), s_2(t), s_3(t), s_4(t), each corresponding to a 2-bit sequence (00,01,10,11)(00, 01, 10, 11).

This increases the data rate per symbol—since each waveform carries qq bits—but requires a more complex receiver to distinguish between the MM signals.

At the receiving end, the transmitted waveform is corrupted by channel effects (e.g., noise, fading, or interference), resulting in a channel-corrupted waveform.

The demodulator processes this received signal and converts each waveform back into a form that estimates the transmitted data symbol.

In binary modulation, the demodulator might output a scalar value (e.g., a voltage level) indicating whether s1(t)s_1(t) or s2(t)s_2(t) was more likely sent.

In MM-ary modulation, it might produce a vector in a signal space (e.g., coordinates in a constellation diagram) representing one of the MM possible symbols.

This output serves as an estimate of the original binary or MM-ary data symbol, though it may still contain errors due to channel noise.

Mathematical Pipeline of the End-to-End Process

Source Input (Information Bits):

The source input is represented as a binary sequence:

u={u1,u2,,uL},ui{0,1}\vec{u} = \{ u_1, u_2, \dots, u_{L} \}, \quad u_i \in \{0, 1\}

This sequence is segmented into Lk\frac{L}{k} blocks, each containing kk bits.

Channel Encoding.

Each kk-bit block ui\vec{u}_i is encoded into an nn-bit codeword ci\vec{c}_i using a generator matrix:

ci=ui×G,ci{0,1}n,i\vec{c}_i = \vec{u}_i \times \mathbf{G}, \quad \vec{c}_i \in \{0, 1\}^n, \quad \forall i

where GF2k×n\mathbf{G} \in \mathbb{F}_2^{k \times n} is the generator matrix over the binary field.

The resulting output bitstream is:

c={c1,c2,,cN},N=Lk×n\vec{c} = \{ c_1, c_2, \dots, c_{N} \}, \quad N = \frac{L}{k} \times n
Bitstream Grouping for Modulation:

The bitstream is grouped into Nq\frac{N}{q} symbols, where each symbol bj\vec{b}_j consists of qq bits:

bj={c(j1)q+1,,cjq},bj{0,1}q\vec{b}_j = \{ c_{(j-1)q+1}, \dots, c_{jq} \}, \quad \vec{b}_j \in \{0, 1\}^q

If N≢0(modq)N \not\equiv 0 \pmod{q}, the bitstream is padded with zeros to ensure its length is divisible by qq.

Symbol Mapping (8-QAM):

Each qq-bit group is mapped to a complex constellation point using the mapping function μ:

sj=μ(bj),μ:{0,1}qC\vec{s}_j = \mu(\vec{b}_j), \quad \mu: \{0,1\}^q \rightarrow \mathbb{C}

where μ assigns qq-bit groups to points in an 8-QAM constellation.

Transmission:

The simplified transmitted signal is constructed as:

x(t)={jsj×ej2πfct}x(t) = \Re\left\{ \sum_j \vec{s}_j \times e^{j2\pi f_c t} \right\}

where fcf_c denotes the carrier frequency, and {}\Re\{\cdot\} represents the real part of the complex expression.

Typically, for passband transmission, x(t)={jsjp(tjTs)ej2πfct}x(t) = \Re\left\{ \sum_j s_j p(t - jT_s) e^{j2\pi f_c t} \right\}, where p(t)p(t) is a pulse-shaping filter and TsT_s is the symbol duration.

Example: (7,4) Linear Block Code with 8-QAM Modulation

System Parameters

Using (7,4) Linear Block Code.

A (7,4) linear block code introduces redundancy to facilitate error correction:

For each 4-bit message, the encoder generates a 7-bit codeword. For example, in systematic encoding:

Input=1011Output=1011xyz\text{Input} = 1011 \rightarrow \text{Output} = 1011xyz

where xyzxyz are parity bits computed from the message bits using the generator matrix.

Bitstream Formation

The encoder processes a sequence of message bits, dividing it into 4-bit blocks.

Each block is encoded into a 7-bit codeword. For example:

These codewords are concatenated into a continuous binary stream:

Bitstream=101110000101101100101\text{Bitstream} = 101110000101101100101

This bitstream is then forwarded to the modulator.

Modulator Input: Preparing for 8-QAM

For 8-QAM:

Mapping Encoded Bits to 8-QAM Symbols

Since the codeword length (7 bits) is not a multiple of 3, the bitstream is treated as continuous and grouped into 3-bit segments:

Bitstream: 101110000101101100101\text{Bitstream: } 101110000101101100101

Grouped into 3-bit segments:

101,110,000,101,101,100,101101, 110, 000, 101, 101, 100, 101

In this example, the total bit count (21 bits) is divisible by 3, so no padding is required. Each 3-bit group is mapped to a unique 8-QAM constellation point, such as:

Each symbol is defined by its amplitude and phase for carrier modulation.

Modulation and Transmission

Each 3-bit group modulates a carrier wave:

Receiver Side (Overview) At the receiver:

Additional Steps

Depending on the system, optional preprocessing or enhancements may be applied between encoding and modulation:

Transmission:

Each symbol is transmitted over the channel using its associated I-Q modulation.

Table 1:Summary of The Pipeline

StageOperation
Encoder(7,4) linear block coding
BitstreamConcatenate 7-bit codewords
ModulatorGroup bits into 3-bit segments
Symbol MappingMap 3-bit groups to 8-QAM constellation
ChannelTransmit analog modulated signal
ReceiverDemodulate + decode

Detection and Decision-Making

The output of the demodulator is fed to the detector, which interprets the estimate and makes a decision regarding the transmitted symbol.

Hard Decision

In the simplest scenario, for binary modulation, the detector determines whether the transmitted bit was a (0)(0) or a (1)(1) based on the scalar output of the demodulator.

For example, if the demodulator provides a value yy and a threshold τ is established (e.g., τ=0\tau = 0 in BPSK), the detector decides:

{y>τ:Bit is 1yτ:Bit is 0\begin{cases} y > \tau: & \text{Bit is } 1 \\ y \leq \tau: & \text{Bit is } 0 \end{cases}

This binary decision is referred to as a hard decision, as it commits definitively to one of two possible outcomes without retaining any ambiguity.

The detection process can be regarded as a form of quantization.

In the hard-decision case, the continuous output of the demodulator (e.g., a real-valued voltage) is quantized into one of two levels, analogous to binary quantization.

The decision boundary (e.g., τ) divides the output space into two regions, each corresponding to a specific bit value.

More generally, the detector can quantize the demodulator output into Q2Q \geq 2 levels, forming a QQ-ary detector. When MM-ary modulation is employed (with M=2qM = 2^q waveforms), the number of quantization levels must satisfy QMQ \geq M to distinguish all possible symbols.

For instance, in QPSK (M=4M = 4), a hard-decision detector might utilize Q=4Q = 4 to map the vector output of the demodulator to one of four symbols.

Soft Decision

In the extreme case, if no quantization is performed (Q=Q = \infty), the detector forwards the unquantized, continuous output directly to the subsequent stage, preserving all information provided by the demodulator.

When Q>MQ > M, the detector offers greater granularity than the number of transmitted symbols, resulting in a soft decision.

For example, in QPSK (M=4M = 4), a detector with Q=8Q = 8 might assign the demodulator output to one of eight levels, providing finer resolution within each symbol’s decision region.

Soft decisions retain more information regarding the likelihood of the received signal rather than enforcing a definitive choice, which can enhance error correction in subsequent decoding processes.

The quantized output of the detector—whether hard or soft—is then passed to the channel decoder.

The decoder leverages the redundancy introduced by the encoder (e.g., the additional bits in each nn-bit codeword) to correct errors induced by channel disturbances.

For hard decisions, the decoder operates with binary or QQ-ary symbols, employing techniques such as Hamming distance minimization.

For soft decisions, it can utilize probabilistic methods, such as maximum likelihood decoding or log-likelihood ratios, to exploit the additional information, typically achieving superior performance in noisy conditions.

Channel Models

A communication channel serves as the medium through which information is transmitted from a sender to a receiver.

The channel’s behavior is mathematically modeled to predict and optimize system performance.

A general communication channel is characterized by three key components:

Probabilistic Channel Model.

These components (X,Y)\bigl(\mathcal{X}, \mathcal{Y}\bigr), and the conditional probability provide a complete probabilistic model of the channel, enabling the analysis of how reliably information can be transmitted.

The channel’s characteristics determine the strategies required for encoding, modulation, and decoding to achieve effective communication.

Memoryless Channels

A channel is classified as memoryless if its output at any given time depends solely on the input at that same time, with no influence from previous inputs or outputs.

Mathematically, a channel is memoryless if the conditional probability of the output sequence y\vec{y} given the input sequence x\vec{x} factors into a product of individual conditional probabilities:

P[yx]=i=1nP[yixi]for all nP[\vec{y} \mid \vec{x}] = \prod_{i=1}^{n} P\bigl[y_i \mid x_i\bigr] \quad \text{for all } n

Here, P[yixi]P[y_i \mid x_i] is the probability of receiving output yiy_i given input xix_i at time index ii, and the product form indicates that each output yiy_i is statistically independent of all other inputs xjx_j (jij \neq i) and outputs yjy_j (jij \neq i), conditioned on xix_i.

This property implies that the channel has no “memory” of past transmissions; the effect of an input xix_i on the output yiy_i is isolated to that specific time instance.

In other words, for a memoryless channel, the output at time ii depends solely on the input at time ii, and the channel’s behavior at each time step is governed by the same conditional probability distribution P[yixi]P[y_i \mid x_i].

This simplifies analysis and design, as the channel can be characterized by a single-symbol transition probability rather than a complex sequence-dependent model.

The simplest and most widely studied memoryless channel model is the binary symmetric channel (BSC).

In the BSC, both the input and output alphabets are binary, i.e., X=Y={0,1}\mathcal{X} = \mathcal{Y} = \{0, 1\}.

The BSC can be defined with the crossover probability pp as:

P[yixi]={1p,if yi=xip,if yixiP[y_i \mid x_i] = \begin{cases} 1 - p, & \text{if } y_i = x_i \\ p, & \text{if } y_i \neq x_i \end{cases}

This model is particularly suitable for systems employing binary modulation (where bits are mapped to two waveforms) and hard decisions at the detector (where the receiver makes a definitive choice between 0 and 1).

The BSC captures the essence of a basic digital communication channel with symmetric error characteristics, making it a foundational concept in information theory and coding.

The Binary Symmetric Channel (BSC) Model

The binary symmetric channel (BSC) model emerges when a communication system is considered as a composite channel, incorporating the modulator, the physical waveform channel, and the demodulator/detector as an integrated unit.

This abstraction is particularly relevant for systems with the following components:

In this setup, the physical channel is modeled as an additive noise channel, where the transmitted waveform is perturbed by random noise (e.g., additive white Gaussian noise, AWGN).

The demodulator and detector together transform the noisy waveform back into a binary sequence.

The resulting composite channel operates in discrete time, with a binary input sequence (from the encoder) and a binary output sequence (from the detector).

This end-to-end system abstracts the continuous-time waveform transmission into a discrete-time model, simplifying analysis.

The BSC model assumes that the combined effects of modulation, channel noise, demodulation, and detection can be represented as a single discrete-time channel with binary inputs and binary outputs.

This abstraction is valid when the noise affects each transmitted bit independently, and the detector’s hard decisions align with the binary nature of the input, making the BSC an appropriate and widely utilized model for such systems.

Characteristics of the Binary Symmetric Channel

The composite channel, modeled as a binary symmetric channel (BSC), is fully characterized by the following:

For the BSC, the channel noise and disturbances are assumed to cause statistically independent errors in the transmitted binary sequence, with an average probability of error (p)(p), known as the crossover probability.

The conditional probabilities are symmetric and defined as:

P[Y=0X=1]=P[Y=1X=0]=p,P[Y=1X=1]=P[Y=0X=0]=1p.\begin{aligned} P[Y = 0 \mid X = 1] &= P[Y = 1 \mid X = 0] = p, \\ P[Y = 1 \mid X = 1] &= P[Y = 0 \mid X = 0] = 1 - p. \end{aligned}

These probabilities can be interpreted as follows:

The symmetry arises because the error probability (p)(p) is identical in both directions (010 \to 1 and 101 \to 0), and the correct reception probability is 1p1 - p.

Since the channel is memoryless, these probabilities apply independently to each transmitted bit, consistent with:

P[yx]=i=1nP[yixi].P[\vec{y} \mid \vec{x}] = \prod_{i=1}^{n} P[y_i \mid x_i].

The BSC is often depicted diagrammatically as a transition model with two inputs and two outputs, connected by arrows labeled with probabilities (p)(p) and (1p)(1 - p).

Channel diagram of the BSC considered above.

Channel diagram of the BSC considered above.

The cascade of the binary modulator, waveform channel, and binary demodulator/detector is thus reduced to this equivalent discrete-time channel, the BSC.

This model simplifies the analysis of error rates and informs the design of error-correcting codes, as (p)(p) (typically 0<p<0.50 < p < 0.5) quantifies the channel’s reliability.

Discrete Memoryless Channels (DMC)

The binary symmetric channel (BSC), discussed previously, is a specific instance of a broader class of channel models known as the discrete memoryless channel (DMC).

A DMC is characterized by two key properties:

A practical example of a DMC arises in a communication system using an MM-ary memoryless modulation scheme.

Here, the modulator maps each input symbol from X\mathcal{X} (with X=M\lvert\mathcal{X}\rvert = M) to one of MM distinct waveforms (e.g., in MM-ary phase-shift keying, M-PSK).

The detector processes the received waveform and produces an output symbol from Y\mathcal{Y}, consisting of QQ-ary symbols (e.g., after hard or soft quantization, where QMQ \geq M, ensures all MM inputs can be distinguished).

The composite channel—comprising the modulator, physical channel, and detector—is thus a DMC, as the modulation and detection processes preserve the discrete and memoryless nature of the system.

The input-output behavior of the DMC is fully described by a set of conditional probabilities P[yx]P[y \mid x], where xXx \in \mathcal{X} and yYy \in \mathcal{Y}.

There are M×QM \times Q such probabilities, one for each possible input-output pair.

For instance, if M=2M = 2 (binary input) and Q=2Q = 2 (binary output), as in the BSC, there are 2×2=42 \times 2 = 4 probabilities (e.g., P[00],P[10],P[01],P[11]P[0\mid0], P[1\mid0], P[0\mid1], P[1\mid1]).

These conditional probabilities can be organized into a probability transition matrix P=[pij]\mathbf{P} = [p_{ij}], where:

The matrix P\mathbf{P} has dimensions X×Y\lvert\mathcal{X}\rvert \times \lvert\mathcal{Y}\rvert (e.g., 2×22 \times 2 for the BSC), and each row sums to 1 (i.e., jpij=1\sum_{j} p_{ij} = 1 for each ii), since these rows represent probability distributions over Y\mathcal{Y} for a given xix_i.

This matrix, often illustrated as in the following figure, provides a compact representation of the DMC’s statistical behavior and facilitates analysis of error rates and channel capacity.

Channel diagram of the considered DMC.

Channel diagram of the considered DMC.

Discrete-Input, Continuous-Output Channels

In contrast to the DMC, the discrete-input, continuous-output channel model relaxes the constraint on the output alphabet while retaining a discrete input.

This model is defined by:

This configuration defines a composite discrete-time memoryless channel, consisting of the modulator, physical channel, and detector.

The channel takes a discrete input XXX \in \mathcal{X} and produces a continuous output YRY \in \mathbb{R}.

Its behavior is characterized by a set of conditional probability density functions (PDFs):

p(yx),xX,yR.p(y \mid x), \quad x \in \mathcal{X}, y \in \mathbb{R}.

For each input symbol xx, p(yx)p(y \mid x) is a PDF over the real line, describing the likelihood of observing a particular output value yy given xx.

Unlike the DMC’s discrete probabilities (P[yx])\bigl(P[y \mid x]\bigr), here p(yx)p(y \mid x) is a continuous function, and the probability of YY falling in an interval [a,b][a, b] is:

abp(yx)dy,\int_{a}^{b} p(y \mid x) \mathrm{d}y,

with:

p(yx)dy=1for each x.\int_{-\infty}^{\infty} p(y \mid x) \mathrm{d}y = 1 \quad \text{for each } x.

This model is relevant when the receiver retains the full resolution of the received signal (e.g., soft-decision outputs) rather than forcing a discrete decision, providing more information for subsequent decoding processes.

Additive White Gaussian Noise (AWGN) Channel

The additive white Gaussian noise (AWGN) channel is one of the most fundamental examples of a discrete-input, continuous-output memoryless channel in communication theory.

Assuming a specific signal mapping, the channel is modeled by:

Y=X+N,Y = X + N,

where:

The term white indicates that the noise has a flat power spectral density (i.e., it is uncorrelated across time), while Gaussian refers to its normal distribution.

For a given input xx, the output YY is a Gaussian random variable with mean xx and variance σ2\sigma^2, thus:

p(yx)=12πσ2exp((yx)22σ2).p(y \mid x) = \frac{1}{\sqrt{2\pi \sigma^2}} \exp \Bigl(- \frac{(y - x)^{2}}{2 \sigma^2}\Bigr).

Multiple Inputs and Outputs

Consider a sequence of nn inputs XiX_i, i=1,2,,ni = 1, 2, \ldots, n. The corresponding outputs are:

Yi=Xi+Ni,i=1,2,,n,Y_i = X_i + N_i, \quad i = 1, 2, \ldots, n,

where each NiN_i is an independent, identically distributed (i.i.d.) Gaussian noise term,

NiN(0,σ2).N_i \sim \mathcal{N}(0, \sigma^2).

Because the channel is memoryless, the noise in each output YiY_i depends only on XiX_i. Formally,

p(y1,y2,,ynx1,x2,,xn)=i=1np(yixi).p(y_1, y_2, \ldots, y_n \big\vert x_1, x_2, \ldots, x_n) = \prod_{i=1}^{n} p(y_i \big\vert x_i).

Substituting the Gaussian PDF yields:

p(y1,y2,,ynx1,x2,,xn)=i=1n12πσ2exp((yixi)22σ2).p(y_1, y_2, \ldots, y_n \big\vert x_1, x_2, \ldots, x_n) = \prod_{i=1}^{n} \frac{1}{\sqrt{2\pi \sigma^2}} \exp \Bigl(- \frac{(y_i - x_i)^{2}}{2 \sigma^2}\Bigr).

This factorization confirms the channel’s memoryless nature, as the joint PDF of the output sequence is the product of individual PDFs, each depending only on the corresponding input.

Role of AWGN Channels

The AWGN channel is a cornerstone of communication theory, providing an accurate model for systems where thermal noise dominates, such as satellite links and wireless channels.

Its importance extends to analyzing modulation schemes (e.g., BPSK, QPSK) with continuous outputs prior to any quantization, forming the basis for many fundamental results in digital communications.

The Discrete-Time AWGN Channel

A discrete-time (continuous-input, continuous-output) additive white Gaussian noise (AWGN) channel is one in which both the input and output take values in the set of all real numbers:

X=Y=R.\mathcal{X} = \mathcal{Y} = \mathbb{R}.

Unlike channels with discrete alphabets, this model permits continuous-valued inputs and outputs, corresponding to a situation with no quantization at either the transmitter or the receiver.

Input–Output Relationship

At each discrete time instant ii, an input xiRx_i \in \mathbb{R} is transmitted over the channel, producing the received symbol:

yi=xi+ni,y_i = x_i + n_i,

where nin_i represents additive noise.

The noise samples {ni}\{n_i\} are independent, identically distributed (i.i.d.) zero-mean Gaussian random variables with variance σ2\sigma^2.

Hence, the PDF of each nin_i is:

pNi(ni)=12πσ2exp(ni22σ2).p_{N_i}(n_i) = \frac{1}{\sqrt{2\pi \sigma^2}} \exp \Bigl(- \frac{n_i^{2}}{2 \sigma^2}\Bigr).

Given an input xix_i, the output yiy_i is a Gaussian random variable with mean xix_i and variance σ2\sigma^2.

Thus, its conditional PDF is:

p(yixi)=12πσ2exp((yixi)22σ2).p(y_i \mid x_i) = \frac{1}{\sqrt{2\pi \sigma^2}} \exp \Bigl(- \frac{\bigl(y_i - x_i\bigr)^{2}}{2 \sigma^2}\Bigr).

Power Constraint

A key practical limitation in this channel model is the power constraint on the input, expressed as an expected power limit:

E[X2]P,\mathbb{E}\bigl[X^{2}\bigr] \le P,

which ensures that the transmitter does not exceed a certain average energy PP. Note that PP represents average power (energy per unit time), not total energy.

For a sequence of nn input symbols:

x=(x1,x2,,xn),\vec{x} = (x_{1}, x_{2}, \ldots, x_{n}),

the time-average power is:

1ni=1nxi2=1nx2,\frac{1}{n} \sum_{i=1}^{n} x_{i}^{2} = \frac{1}{n} \|\vec{x}\|^{2},

where:

x2=i=1nxi2\|\vec{x}\|^{2} = \sum_{i=1}^{n} x_{i}^{2}

is the squared Euclidean norm of x\vec{x}.

As nn grows large, the law of large numbers implies that, with high probability, the time-average power 1nx2\frac{1}{n}\|\vec{x}\|^{2} converges to E[X2]\mathbb{E}[X^2].

Thus, the constraint:

1nx2P\frac{1}{n} \|\vec{x}\|^{2} \le P

arises naturally.

In simpler terms:

i=1nxi2nP.\sum_{i=1}^{n} x_{i}^{2} \le n P.

Geometric Interpretation

Geometrically, the set of all allowable input sequences x\vec{x} lies within an nn-dimensional sphere of radius nP\sqrt{n P} centered at the origin, since:

(nP)2=nPi=1nxi2nP.(\sqrt{n P})^{2} = n P \quad\Longleftrightarrow\quad \sum_{i=1}^{n} x_{i}^{2} \le n P.

This spherical boundary in nn-dimensional space is crucial for understanding both the channel capacity and the design of signal constellations under energy constraints.

The AWGN Waveform Channel

The AWGN waveform channel describes a physical communication medium in which both the input and output are continuous-time waveforms, rather than discrete symbols.

This can be interpreted as a continuous-time, continuous-input, continuous-output AWGN channel.

To highlight the core behavior of the physical channel, the modulator and demodulator are treated as separate from the channel model, directing attention solely to the process of waveform transmission.

Suppose the channel has a bandwidth WW, characterized by an ideal frequency response:

C(f)=1forfWC(f) = 1 \quad \text{for} \quad |f| \leq W

and

C(f)=0otherwise.C(f) = 0 \quad \text{otherwise}.

This indicates that the channel perfectly transmits signals whose frequency components lie in the interval [W,+W][-W, +W] and suppresses those outside this range.

The input waveform x(t)x(t) is assumed to be band-limited, such that its Fourier transform satisfies:

X(f)=0forf>W,X(f) = 0 \quad \text{for} \quad |f| > W,

ensuring conformity with the channel’s bandwidth.

At the channel output, the waveform y(t)y(t) is given by:

y(t)=x(t)+n(t),y(t) = x(t) + n(t),

where n(t)n(t) is a sample function of an additive white Gaussian noise (AWGN) process.

The noise has a power spectral density:

N02(W/Hz),\frac{N_0}{2} \quad \text{(W/Hz)},

indicating that its power is distributed uniformly across all frequencies.

For a channel of bandwidth WW, the noise power confined within the interval [W,+W][-W, +W] is:

σ2=WWN02df=N02×2W=N0W.\sigma^2 = \int_{-W}^{W} \frac{N_0}{2} df = \frac{N_0}{2} \times 2W = N_0 W.

As will be clarified later, the discrete-time equivalent of this channel provides a simpler perspective through sampling.

Power Constraint and Signal Representation

The input waveform x(t)x(t) must obey a power constraint:

E[x2(t)]P,\mathbb{E}[x^2(t)] \leq P,

which restricts the expected instantaneous power of x(t)x(t) to PP.

For ergodic processes, where time averages equal ensemble averages (as is the case for stationary processes), this is expressed as:

limT1TT/2T/2x2(t)dtP.\lim_{T \to \infty} \frac{1}{T} \int_{-T/2}^{T/2} x^2(t) dt \leq P.

Interpreted over an interval of length TT, this stipulates that the average energy per unit time cannot exceed PP.

Consequently, this condition aligns with that represented via E[x2(t)]\mathbb{E}[x^2(t)].

To analyze the channel in probabilistic terms, x(t)x(t), y(t)y(t), and n(t)n(t) are expanded in terms of a complete set of orthonormal functions {ϕj(t)}\{\phi_j(t)\}.

When a signal has bandwidth WW and duration TT, its dimension in signal space can be approximated by 2WT2WT.

This approximation follows from the sampling theorem:

Thus, the signal space effectively has 2W2W dimensions per second.

Orthonormal Expansion

Using this orthonormal set, the waveforms can be written as:

x(t)=jxjϕj(t),x(t) = \sum_{j} x_j \phi_j(t),
n(t)=jnjϕj(t),n(t) = \sum_{j} n_j \phi_j(t),
y(t)=jyjϕj(t),y(t) = \sum_{j} y_j \phi_j(t),

where {ϕj(t),j=1,2,,2WT}\{\phi_j(t), j = 1, 2, \ldots, 2WT\} are orthonormal basis functions (e.g., sinc functions or prolate spheroidal wave functions) satisfying:

ϕi(t)ϕj(t)dt=δij={1,if i=j,0,if ij.\int \phi_i(t) \phi_j(t) dt = \delta_{ij} = \begin{cases} 1, & \text{if } i = j, \\ 0, & \text{if } i \neq j. \end{cases}

The expansion coefficients are:

xj=x(t)ϕj(t)dt,nj=n(t)ϕj(t)dt,yj=y(t)ϕj(t)dt,x_j = \int x(t) \phi_j(t) dt, \quad n_j = \int n(t) \phi_j(t) dt, \quad y_j = \int y(t) \phi_j(t) dt,

representing the projections of the signals onto these basis functions.

Since y(t)=x(t)+n(t)y(t) = x(t) + n(t), substituting the expansions into this relationship results in:

jyjϕj(t)=jxjϕj(t)+jnjϕj(t).\sum_{j} y_j \phi_j(t) = \sum_{j} x_j \phi_j(t) + \sum_{j} n_j \phi_j(t).

By orthonormality, matching coefficients across the sums yields:

yj=xj+nj.y_j = x_j + n_j.

Because n(t)n(t) is white Gaussian noise with power spectral density N02\frac{N_0}{2}, the noise coefficients njn_j are independent and identically distributed (i.i.d.) Gaussian random variables with zero mean and variance σ2=N02\sigma^2 = \frac{N_0}{2}.

Hence, each dimension of the expansion carries a noise variance of N02\frac{N_0}{2}, consistent with the total noise power spread over the channel’s bandwidth.

Equivalent Discrete-Time Channel

The AWGN waveform channel can be reduced to a discrete-time model in which each output coefficient yjy_j is related to the corresponding input coefficient xjx_j through:

yj=xj+nj.y_j = x_j + n_j.

The conditional probability density function (PDF) for each output symbol given the input symbol is:

p(yjxj)=12πσ2exp((yjxj)22σ2)=1πN0exp((yjxj)2N0),\begin{split} p(y_j \mid x_j) &= \frac{1}{\sqrt{2\pi \sigma^2}} \exp \Bigl(-\frac{(y_j - x_j)^2}{2\sigma^2}\Bigr) \\ &= \frac{1}{\sqrt{\pi N_0}} \exp \Bigl(-\frac{(y_j - x_j)^2}{N_0}\Bigr), \end{split}

because:

σ2=N02and2πσ2=2π×N02=πN0.\sigma^2 = \frac{N_0}{2} \quad \text{and} \quad \sqrt{2\pi \sigma^2} = \sqrt{2\pi \times \frac{N_0}{2}} = \sqrt{\pi N_0}.

Since the noise coefficients njn_j are independent for different values of jj, the overall channel is memoryless, which gives:

p(y1,y2,,yNx1,x2,,xN)=j=1Np(yjxj).p(y_1, y_2, \ldots, y_N \mid x_1, x_2, \ldots, x_N) = \prod_{j=1}^{N} p(y_j \mid x_j).

Vector AWGN Model

From the relationship:

yj=xj+nj,with njN(0,σ2=N0/2)for j=1,2,,N,y_j = x_j + n_j, \quad \text{with } n_j \sim \mathcal{N}(0, \sigma^2 = N_0/2) \quad \text{for } j = 1, 2, \dots, N,

this can be rewritten in a compact vector form:

y=x+n,nN(0,N02I)\boxed{\vec{y} = \vec{x} + \vec{n}, \quad \vec{n} \sim \mathcal{N}(\vec{0}, \frac{N_0}{2} \mathbf{I})}

where:

Power Constraint and Parseval’s Theorem

The continuous-time power constraint translates directly to the discrete coefficients.

By Parseval’s theorem, for a signal of duration TT:

1TT/2T/2x2(t)dt=1Tj=12WTxj2.\frac{1}{T} \int_{-T/2}^{T/2} x^2(t) dt = \frac{1}{T} \sum_{j=1}^{2WT} x_j^2.

In this interval of length TT, there are 2WT2WT coefficients, so the average power per coefficient is:

E[X2]=12WTj=12WTE[Xj2].\mathbb{E}[X^2] = \frac{1}{2WT} \sum_{j=1}^{2WT} \mathbb{E}[X_j^2].

Hence:

limT1TT/2T/2x2(t)dt=limT1Tj=12WTxj2=2WE[X2]P.\lim_{T \to \infty} \frac{1}{T} \int_{-T/2}^{T/2} x^2(t) dt = \lim_{T \to \infty} \frac{1}{T} \sum_{j=1}^{2WT} x_j^2 = 2W \mathbb{E}[X^2] \leq P.

Solving for E[X2]\mathbb{E}[X^2], one obtains:

E[X2]P2W.\mathbb{E}[X^2] \leq \frac{P}{2W}.

Accordingly, a waveform channel of bandwidth WW and input power PP behaves like 2W2W uses per second of a discrete-time AWGN channel whose noise variance is σ2=N02\sigma^2 = \frac{N_0}{2}.

This equivalence establishes the connection between the continuous-time channel and its discrete-time counterpart.