Signals & Systems — From Time Domain to Spectral Thinking

1. What Signals and Systems Theory Actually Does

Every piece of technology that senses, communicates, or controls involves signals (functions that carry information) and systems (operators that transform those signals). A microphone converts air pressure into a voltage signal; an amplifier system scales it; a filter system removes hum; an ADC system samples it; an FFT algorithm reveals its frequency content; a codec compresses it; a DAC reconstructs it; a speaker converts it back to pressure.

Signals and Systems theory provides the mathematical machinery to predict, analyze, and design every stage of this chain. The core insight is that certain classes of systems — Linear Time-Invariant (LTI) — are completely characterized by a single function (the impulse response $h(t)$ or $h[n]$), and that transforming from time to frequency domain converts the hard operation (convolution) into the easy one (multiplication).


2. Signal Fundamentals

2.1 Continuous-Time (CT) vs. Discrete-Time (DT)

DT signals are not inherently “digital” — they are just defined at discrete instants. Digital signals are further quantized in amplitude.

2.2 Energy and Power

This distinction matters because the Fourier Transform exists for energy signals, while periodic (power) signals need the Fourier Series or generalized functions ($\delta$-functions in the spectrum).

2.3 Elementary Signals — The Building Blocks

Unit Impulse $\delta(t)$ / $\delta[n]$: The single most important signal. In CT, $\delta(t)$ is not a function in the usual sense — it is a distribution defined by its action under integration: \(\int_{-\infty}^{\infty} f(t)\delta(t-t_0) dt = f(t_0)\) This is the sifting property. It says: the impulse “picks out” the value of $f$ at $t_0$. Why does this matter? Because any signal can be decomposed as a weighted sum of shifted impulses: \(x(t) = \int_{-\infty}^{\infty} x(\tau)\delta(t-\tau) d\tau\) In DT, $\delta[n]$ is simply 1 at $n=0$ and 0 everywhere else — no distribution subtlety needed.

Unit Step $u(t)$ / $u[n]$: $u(t) = 1$ for $t \geq 0$, $0$ for $t < 0$. Related to the impulse by $u(t) = \int_{-\infty}^{t}\delta(\tau)d\tau$, or equivalently $\delta(t) = \frac{du}{dt}$. The step is the canonical “switching on” signal — it models turning on a voltage, opening a valve, starting a motor.

Complex Exponential $e^{st}$ / $z^n$: The eigenfunctions of LTI systems. If you feed $e^{st}$ into an LTI system, you get $H(s)e^{st}$ out — the system scales and phase-shifts, but doesn’t change the functional form. This is why the Laplace and Z transforms work: they decompose arbitrary signals into sums of exponentials, and each exponential passes through the system independently.


3. LTI Systems: The Framework

3.1 Why LTI?

Linearity (superposition + scaling) means: if input $x_1 \to y_1$ and $x_2 \to y_2$, then $ax_1 + bx_2 \to ay_1 + by_2$. This is incredibly powerful — it means we can analyze each component of a signal independently and add the results.

Time Invariance means: if $x(t) \to y(t)$, then $x(t - t_0) \to y(t - t_0)$. The system behaves the same way regardless of when the input arrives. A guitar amp that distorts is nonlinear. A time-varying channel (fading in wireless) is not time-invariant.

Together, LTI means the system is completely characterized by its response to a single impulse.

3.2 Impulse Response and Convolution

If the system’s response to $\delta(t)$ is $h(t)$, then the response to any input $x(t)$ is:

\[y(t) = x(t) * h(t) = \int_{-\infty}^{\infty} x(\tau) h(t - \tau) d\tau\]

Why this works: We decomposed $x(t)$ into infinitely many shifted, scaled impulses ($x(\tau)\delta(t-\tau)d\tau$). By linearity, the response is the sum of the system’s responses to each impulse. By time invariance, the response to $\delta(t-\tau)$ is $h(t-\tau)$. Add them all up → convolution integral.

In DT, the convolution sum: \(y[n] = \sum_{k=-\infty}^{\infty} x[k] \, h[n-k]\)

Graphical convolution (flip $h$, slide it across $x$, compute the overlap area at each position) builds powerful intuition for what convolution does: it smears, smooths, or shapes the input according to the impulse response.

3.3 System Properties from $h(t)$


4. Time-Domain Analysis of CT Systems

4.1 Differential Equation Representation

CT LTI systems are described by constant-coefficient linear ODEs: \(\sum_{k=0}^{N} a_k \frac{d^k y(t)}{dt^k} = \sum_{m=0}^{M} b_m \frac{d^m x(t)}{dt^m}\)

For example, a simple RC lowpass filter: $RC\dot{y} + y = x$. This is a 1st-order ODE with $a_0 = 1$, $a_1 = RC$, $b_0 = 1$.

4.2 Impulse Response of CT Systems

To find $h(t)$, set $x(t) = \delta(t)$ and solve the ODE with zero initial conditions. For the RC filter: \(h(t) = \frac{1}{RC} e^{-t/(RC)} u(t)\) This is an exponential decay. The time constant $\tau = RC$ controls how fast the system “forgets” — after $5\tau$, the response has decayed to <1% of its peak. A large $\tau$ means the system averages over a long time window (heavy smoothing); a small $\tau$ means it reacts quickly (light smoothing).

4.3 Step Response

The step response $s(t)$ is the response to $u(t)$: \(s(t) = \int_{-\infty}^{t} h(\tau) d\tau = h(t) * u(t)\) For the RC filter: $s(t) = (1 - e^{-t/(RC)}) u(t)$. It starts at 0 and exponentially approaches 1. The step response reveals:

4.4 Second-Order Systems

\(\ddot{y} + 2\zeta\omega_n \dot{y} + \omega_n^2 y = \omega_n^2 x\) Characterized by:

The step response of an underdamped system exhibits overshoot $\approx e^{-\pi\zeta/\sqrt{1-\zeta^2}} \times 100\%$ and settling time $\approx 4/(\zeta\omega_n)$. These formulas drive control system design — choosing $\zeta$ and $\omega_n$ sets the speed-vs-overshoot tradeoff.


5. Time-Domain Analysis of DT Systems

5.1 Difference Equation Representation

DT LTI systems are described by constant-coefficient linear difference equations: \(\sum_{k=0}^{N} a_k y[n-k] = \sum_{m=0}^{M} b_m x[n-m]\)

For example, a first-order DT lowpass: $y[n] = \alpha y[n-1] + (1-\alpha) x[n]$. This is a single-pole recursive filter — the DT analog of the RC circuit.

5.2 DT Impulse and Step Response

Set $x[n] = \delta[n]$: the impulse response $h[n]$ is obtained by iterating the difference equation. For $y[n] = \alpha y[n-1] + (1-\alpha)\delta[n]$: \(h[n] = (1-\alpha)\alpha^n u[n]\) If $|\alpha| < 1$, $h[n]$ decays geometrically → stable. If $|\alpha| \geq 1$, $h[n]$ grows → unstable.

The DT step response is: $s[n] = \sum_{k=0}^{n} h[k] = (1 - \alpha^{n+1}) u[n]$. It approaches 1 as $n \to \infty$ (for $ \alpha < 1$), just like the CT RC step response approaches 1 as $t \to \infty$.

5.3 The Moving Average Filter

\(y[n] = \frac{1}{M} \sum_{k=0}^{M-1} x[n-k]\) This is an FIR (Finite Impulse Response) system: $h[n]$ has only $M$ non-zero terms. It averages the last $M$ samples — a simple smoother. Its frequency response is a sinc-like function with nulls at multiples of $2\pi/M$. Wider window ($M$ larger) → narrower passband → more smoothing.


6. Frequency-Domain Analysis: CT (Fourier & Laplace)

6.1 Fourier Series (Periodic CT Signals)

A periodic signal $x(t)$ with period $T_0$ can be decomposed into harmonics: \(x(t) = \sum_{k=-\infty}^{\infty} C_k \, e^{jk\omega_0 t}, \quad \omega_0 = \frac{2\pi}{T_0}\) \(C_k = \frac{1}{T_0} \int_{T_0} x(t) e^{-jk\omega_0 t} dt\)

What this means: Any periodic signal — a square wave, a sawtooth, a heartbeat — is a superposition of sinusoids at integer multiples of the fundamental frequency $\omega_0$. The coefficients $C_k$ tell you how much energy is at each harmonic. A square wave has strong odd harmonics ($C_1, C_3, C_5, …$) falling off as $1/k$ — this is why a square wave sounds “buzzy” compared to a pure sine (which has only $C_1$).

Gibbs phenomenon: Truncating the series at $K$ terms produces ~9% overshoot at discontinuities, regardless of $K$. The overshoot doesn’t shrink — it just gets narrower. This is a fundamental limit of representing a discontinuity with a finite number of smooth sinusoids.

6.2 Continuous-Time Fourier Transform (CTFT)

For aperiodic signals: \(X(j\omega) = \int_{-\infty}^{\infty} x(t) e^{-j\omega t} dt\) \(x(t) = \frac{1}{2\pi} \int_{-\infty}^{\infty} X(j\omega) e^{j\omega t} d\omega\)

$X(j\omega)$ is the spectrum — it tells you the density of signal content at each frequency $\omega$. The magnitude $ X(j\omega) $ is the amplitude spectrum; the angle $\angle X(j\omega)$ is the phase spectrum.

Key transform pairs (build your intuition around these):

Key properties:

6.3 Laplace Transform

The Fourier Transform only exists for signals whose integral converges (energy signals, or with generalized functions). The Laplace Transform generalizes by adding a convergence factor: \(X(s) = \int_{0}^{\infty} x(t) e^{-st} dt, \quad s = \sigma + j\omega\)

Setting $\sigma = 0$ recovers the Fourier Transform (on the imaginary axis). The extra degree of freedom $\sigma$ allows us to handle growing signals (like $e^{2t}$) and, critically, to solve ODEs with initial conditions.

Why the s-domain is powerful for system analysis:

Pole-zero interpretation:

Region of Convergence (ROC): The set of $s$ values where the Laplace integral converges. For causal signals, ROC is to the right of the rightmost pole. The ROC determines whether a given pole-zero map corresponds to a causal/stable, causal/unstable, or anticausal system.

6.4 Partial Fraction Expansion

To find $h(t)$ from $H(s)$, decompose $H(s)$ into simple fractions: \(H(s) = \frac{A_1}{s - p_1} + \frac{A_2}{s - p_2} + \cdots\) Each term inverts to $A_i e^{p_i t} u(t)$ (for causal systems). This is how we go from an algebraic transfer function back to a time-domain impulse response.


7. Frequency-Domain Analysis: DT (DTFT & Z-Transform)

7.1 Discrete-Time Fourier Transform (DTFT)

\(X(e^{j\Omega}) = \sum_{n=-\infty}^{\infty} x[n] e^{-j\Omega n}\) \(x[n] = \frac{1}{2\pi} \int_{-\pi}^{\pi} X(e^{j\Omega}) e^{j\Omega n} d\Omega\)

The key difference from CTFT: the DTFT is periodic in $\Omega$ with period $2\pi$. This is because $e^{j(\Omega + 2\pi)n} = e^{j\Omega n}$ for integer $n$. Frequencies $\Omega$ and $\Omega + 2\pi$ are indistinguishable in DT — this is the mathematical root of aliasing.

The unique frequency range is $-\pi < \Omega \leq \pi$ (or equivalently $0$ to $2\pi$). $\Omega = 0$ is DC; $\Omega = \pi$ is the highest frequency representable (Nyquist frequency), corresponding to the pattern $+1, -1, +1, -1, \ldots$

7.2 Z-Transform

\[X(z) = \sum_{n=-\infty}^{\infty} x[n] z^{-n}, \quad z \in \mathbb{C}\]

The Z-Transform is to DT what the Laplace Transform is to CT. Setting $z = e^{j\Omega}$ (the unit circle) recovers the DTFT. The extra freedom of complex $z$ allows us to handle growing sequences and solve difference equations.

Transfer function: \(H(z) = \frac{Y(z)}{X(z)} = \frac{b_0 + b_1 z^{-1} + \cdots + b_M z^{-M}}{1 + a_1 z^{-1} + \cdots + a_N z^{-N}}\)

Geometric interpretation: At any frequency $\Omega$, $ H(e^{j\Omega}) $ is proportional to the product of distances from zeros to the point $e^{j\Omega}$ divided by the product of distances from poles. A pole near the unit circle creates a peak in the frequency response (resonance); a zero near the unit circle creates a dip (notch).

7.3 DFT and FFT

The Discrete Fourier Transform (DFT) is a sampled version of the DTFT for finite-length sequences: \(X[k] = \sum_{n=0}^{N-1} x[n] e^{-j2\pi kn/N}, \quad k = 0, 1, \ldots, N-1\)

It transforms $N$ time samples into $N$ frequency samples. The FFT (Fast Fourier Transform) computes the DFT in $O(N \log N)$ instead of $O(N^2)$ by exploiting symmetry (Cooley-Tukey). This makes real-time spectral analysis, fast convolution, and compressed sensing practical.

Spectral leakage: The DFT assumes the signal is periodic with period $N$. If the signal’s period doesn’t divide $N$ evenly, the DFT “sees” a discontinuity at the boundary, spreading energy across all bins. Windowing (Hamming, Hanning, Blackman) smoothly tapers the signal edges to reduce leakage at the cost of wider main lobes (reduced frequency resolution).


8. Sampling and Reconstruction

8.1 The Sampling Theorem (Nyquist-Shannon)

Ideal sampling multiplies $x(t)$ by a Dirac comb: \(x_s(t) = x(t) \sum_{k=-\infty}^{\infty} \delta(t - kT_s)\)

In the frequency domain, this replicates the spectrum at multiples of $\omega_s = 2\pi/T_s$: \(X_s(j\omega) = \frac{1}{T_s} \sum_{k=-\infty}^{\infty} X(j(\omega - k\omega_s))\)

If $\omega_s > 2\omega_{max}$ (the Nyquist criterion), the replicas don’t overlap, and the original spectrum can be perfectly recovered by an ideal lowpass filter with cutoff $\omega_s/2$. If $\omega_s \leq 2\omega_{max}$, the replicas overlap (aliasing), and the original signal is irrecoverably corrupted.

Practical implications:

8.2 Reconstruction

Ideal reconstruction convolves the samples with a sinc function: \(x(t) = \sum_{n} x[n] \, \text{sinc}\left(\frac{t - nT_s}{T_s}\right)\) Each sample contributes a sinc pulse, and the sum reconstructs $x(t)$ exactly (if Nyquist was satisfied). But a sinc is non-causal and infinite — impractical.

Zero-Order Hold (ZOH): The simplest practical DAC — hold each sample value constant until the next sample. The ZOH output has a staircase shape. Its frequency response is $H_{ZOH}(j\omega) = T_s \,\text{sinc}(\omega T_s / 2\pi) \, e^{-j\omega T_s/2}$, which introduces a sinc droop in the passband. Compensation (“sinc correction”) flattens this.


9. Analog (CT) Filter Design

Filters are defined by what they pass and what they stop.

9.1 Ideal vs. Realizable

An ideal lowpass filter has $H(j\omega) = 1$ for $ \omega < \omega_c$ and $0$ otherwise. Its impulse response is $h(t) = \text{sinc}(\omega_c t / \pi)$ — non-causal (extends to $t = -\infty$) and infinite in duration. No physical system can implement this. All real filter design is about approximating this ideal within constraints.

9.2 Filter Specifications

9.3 Classic Analog Filter Families

Butterworth (Maximally Flat): \(|H(j\omega)|^2 = \frac{1}{1 + (\omega/\omega_c)^{2N}}\)

Chebyshev Type I (Equiripple Passband):

Chebyshev Type II (Equiripple Stopband):

Elliptic (Cauer):

Bessel (Maximally Flat Group Delay):

9.4 Filter Transformations

All the above are designed as lowpass prototypes. Standard transformations convert them:

This means you design a normalized lowpass, then algebraically transform it to any type.


10. Digital (DT) Filter Design

10.1 FIR Filters (Finite Impulse Response)

An FIR filter has the form: \(y[n] = \sum_{k=0}^{M} b_k x[n-k]\)

There are no feedback terms (no $a_k y[n-k]$). The impulse response IS the coefficient vector ${b_0, b_1, \ldots, b_M}$ — it lasts exactly $M+1$ samples and then dies.

Advantages:

Disadvantages:

Design methods:

10.2 IIR Filters (Infinite Impulse Response)

An IIR filter has the form: \(y[n] = \sum_{m=0}^{M} b_m x[n-m] - \sum_{k=1}^{N} a_k y[n-k]\)

The feedback terms ($a_k y[n-k]$) create poles, making the impulse response theoretically infinite in duration (though it decays if the filter is stable).

Advantages:

Disadvantages:

Design methods:

10.3 FIR vs. IIR: When to Use Which

Criterion FIR IIR
Stability Guaranteed Must verify
Linear phase Easy (symmetric) Not possible (causal)
Order for sharp cutoff High (50–200+) Low (4–12)
Latency Higher (long filter) Lower
Coefficient sensitivity Low High (poles near unit circle)
Adaptive filtering Preferred (LMS, RLS) Possible but tricky

Rule of thumb: Use FIR when linear phase or guaranteed stability matters (audio, communications, medical). Use IIR when computational budget is tight and some phase distortion is acceptable (control systems, real-time embedded).


11. Digital Signal Processing (DSP) Implementation

11.1 Fixed-Point vs. Floating-Point

Quantization noise: When a continuous signal is converted to $B$-bit fixed-point, each sample has a rounding error uniformly distributed in $[-\Delta/2, +\Delta/2]$ where $\Delta = 2^{-B} \times \text{full scale}$. The Signal-to-Quantization-Noise Ratio is approximately: \(\text{SQNR} \approx 6.02B + 1.76 \text{ dB}\) Each additional bit adds ~6 dB. A 16-bit audio ADC achieves ~98 dB SQNR; a 24-bit ADC achieves ~146 dB.

11.2 Filter Structures

The same transfer function $H(z)$ can be implemented in multiple structures with different numerical properties:

11.3 Multirate DSP


12. Modulation and Communication Systems

12.1 Why Modulate?

Baseband signals (audio: 20 Hz–20 kHz, video: 0–6 MHz) cannot be transmitted directly over wireless channels — the antenna would need to be hundreds of kilometers long ($\lambda/4$ at 1 kHz). Modulation shifts the signal to a carrier frequency $f_c$ where efficient transmission is possible.

12.2 Amplitude Modulation (AM)

\(y(t) = [1 + m \cdot x(t)] \cos(\omega_c t)\) The spectrum of $x(t)$ is shifted to $\pm \omega_c$. Simple demodulation (envelope detection), but wastes power and bandwidth (carrier + two sidebands).

12.3 Frequency Modulation (FM)

The instantaneous frequency varies with the signal: $\omega_i(t) = \omega_c + k_f x(t)$. FM is more resistant to amplitude noise than AM — the information is in the frequency, so amplitude variations (static, interference) can be clipped without losing the signal. This is why FM radio sounds cleaner than AM.

12.4 Digital Modulation


13. Stochastic (Random) Signals and Processes

Real-world signals are rarely deterministic — sensor noise, channel fading, stock prices, and biological signals all have random components. Stochastic process theory provides the tools to analyze systems driven by random inputs.

13.1 Random Variables and Processes

A random process $X(t)$ (or $X[n]$) is a collection of random variables indexed by time. At each fixed $t$, $X(t)$ is a random variable with a probability distribution. One realization (one experiment) gives a deterministic function of time — a sample function.

13.2 Stationarity and Ergodicity

13.3 Power Spectral Density (PSD)

For a WSS process, the Wiener-Khinchin theorem says: \(S_{XX}(\omega) = \mathcal{F}\{R_{XX}(\tau)\} = \int_{-\infty}^{\infty} R_{XX}(\tau) e^{-j\omega\tau} d\tau\) The PSD $S_{XX}(\omega)$ tells you how the signal’s power is distributed across frequency.

White noise: $S_{NN}(\omega) = N_0/2$ (constant across all frequencies). The autocorrelation of white noise is $R_{NN}(\tau) = \frac{N_0}{2}\delta(\tau)$ — samples are uncorrelated at any nonzero lag. White noise is the standard model for thermal noise, quantization noise, and many interference sources.

13.4 LTI Systems with Random Inputs

If WSS input $X(t)$ with PSD $S_{XX}(\omega)$ is fed through an LTI system with transfer function $H(j\omega)$, the output is also WSS with: \(S_{YY}(\omega) = |H(j\omega)|^2 S_{XX}(\omega)\) This is enormously useful:

13.5 Noise in Practical Systems


14. Time-Frequency Analysis

14.1 The Uncertainty Principle

You cannot simultaneously have perfect time resolution and perfect frequency resolution. For any signal: \(\Delta t \cdot \Delta \omega \geq \frac{1}{2}\) A short pulse is well-localized in time but spread in frequency. A pure sinusoid is perfectly localized in frequency but extends forever in time. The Gaussian achieves equality.

14.2 Short-Time Fourier Transform (STFT)

\(\text{STFT}\{x(t)\}(\tau, \omega) = \int x(t) w(t - \tau) e^{-j\omega t} dt\) Chop the signal into overlapping windows $w(t-\tau)$, compute the Fourier Transform of each chunk. The result is a spectrogram — a 2D plot of frequency content vs. time. Short windows → good time resolution, poor frequency resolution. Long windows → vice versa. The window length is fixed, so STFT has uniform time-frequency resolution.

14.3 Wavelet Transform

Uses scaled and shifted versions of a mother wavelet $\psi(t)$: \(W(a,b) = \frac{1}{\sqrt{|a|}} \int x(t) \psi^*\left(\frac{t-b}{a}\right) dt\) At small scale $a$: short, high-frequency wavelet → good time resolution for transients. At large scale $a$: long, low-frequency wavelet → good frequency resolution for slowly varying components. This multi-resolution behavior makes wavelets ideal for signals with both sharp events and sustained tones (speech, seismic data, ECG).


15. The Unified View

All of signals and systems theory rests on a few deep connections:

Concept CT DT
Elementary signal $e^{st}$ $z^n$
Transform Laplace $\mathcal{L}$ Z-Transform $\mathcal{Z}$
Frequency analysis CTFT ($s = j\omega$) DTFT ($z = e^{j\Omega}$)
Periodic signals Fourier Series Discrete Fourier Series
Stability region Left-half $s$-plane Inside unit circle in $z$
Convolution $\int x(\tau)h(t-\tau)d\tau$ $\sum x[k]h[n-k]$
System description ODE Difference equation
Filter design Butterworth, Cheby, Elliptic FIR (Parks-McClellan), IIR (Bilinear)

The Laplace Transform is the mother transform for CT, and the Z-Transform is the mother transform for DT. The Fourier Transform is a special case of each (evaluate on the boundary of the stability region). Sampling bridges the two worlds: it converts a CT signal into a DT signal, and the Z-Transform of the sampled signal relates to the Laplace Transform of the original via $z = e^{sT_s}$.

Understanding this architecture lets you move fluidly between domains and choose the right tool for each problem.