Control Theory Basics

1. Why Control Starts with Dynamical Systems

Control theory is not mainly about tuning gains. It begins with a more basic question:

What kind of system are we trying to influence, and how does it move on its own before we intervene?

That is why control is inseparable from dynamics.

A heater does not jump instantly to the commanded temperature because heat must accumulate in thermal mass. A positioning stage does not move instantly to the desired location because mass, damping, and elasticity all matter. A motor does not reject load disturbances automatically because its natural dynamics do not know what target we care about.

So the logic of the subject is:

  1. describe the system,
  2. build a model of its dynamics,
  3. analyze how that model behaves,
  4. decide whether the behavior is stable and acceptable,
  5. then design feedback to reshape the behavior.

If Signals & Systems gives the language of LTI models, control theory gives that language a purpose: shaping motion, regulation, and robustness in the presence of disturbances and uncertainty.


2. What a System Model Is

2.1 System, plant, input, output, and disturbance

A system is any dynamical object whose future behavior depends on its current condition and on what acts on it.

In control, we usually separate that system into the following roles:

This separation matters because control is not “change the system however you like.” It is “use the available input channel to influence the measured output despite disturbances.”

2.2 State, parameters, and operating point

A state is the smallest set of variables that, together with future inputs, determines future behavior.

Examples:

Parameters such as mass, damping, stiffness, resistance, heat capacity, and time constants are not states. They do not evolve because of the current motion in the same way; they describe the plant itself.

An operating point is an equilibrium around which we choose to study or control the system. Real plants are often nonlinear, and a good model is usually a good model around a chosen operating regime, not everywhere in the universe.

2.3 What a model actually does

A model is not the plant. A model is a controlled simplification that keeps the dominant physics and discards the detail that does not matter for the question at hand.

Good control models usually answer these three questions:

  1. what variables store energy or memory,
  2. what inputs and disturbances drive those variables,
  3. what output the controller will see.

The model is “good” if it is simple enough to analyze and accurate enough to support design.

2.4 Modelling assumptions are part of the model

When we write a control model, we are also silently making assumptions:

Those assumptions are not optional footnotes. They define where the control design can be trusted.


3. From Physics to Equations

To make the modelling idea concrete, keep two running examples in mind:

3.1 First-order example: a heated chamber

Let $\theta(t)$ denote the temperature rise above ambient. A simple energy balance gives

\[C \dot{\theta} = -\frac{1}{R}\theta + K_h u + d,\]

where:

This is commonly rewritten as

\[\tau \dot{\theta} + \theta = K u + d_\theta,\]

with $\tau = RC$.

What this means physically is simple: the chamber has one dominant stored-energy state and therefore one dominant time scale. If you stop forcing it, it forgets exponentially rather than instantly.

3.2 Second-order example: a position servo

For a mass-spring-damper system driven by force input $u(t)$,

\[m \ddot{x} + b \dot{x} + kx = u + d.\]

This model already contains the basic ingredients of servo dynamics:

What this means physically is that the plant can oscillate. The system does not merely “lag”; it can overshoot, ring, and store kinetic and potential energy.

3.3 Input-output and state-space forms

The thermal model is naturally written as an input-output ODE. The mechanical model can also be written that way, but a state-space form makes the internal memory explicit:

\[\dot{x}_1 = x_2, \qquad \dot{x}_2 = -\frac{k}{m}x_1 - \frac{b}{m}x_2 + \frac{1}{m}u + \frac{1}{m}d,\]

with output

\[y = x_1.\]

In matrix form,

\[\dot{x} = Ax + Bu + Ed, \qquad y = Cx.\]

Input-output models are often enough for classical loop design. State-space models are better when we care about internal modes, multiple states, multiple inputs, or structured modern design.

3.4 Nonlinear models and linearization

A general nonlinear plant may be written as

\[\dot{x} = f(x,u,d), \qquad y = h(x,u,d).\]

Around an equilibrium $(x_0,u_0,d_0)$ satisfying

\[f(x_0,u_0,d_0) = 0,\]

small deviations obey the linearized model

\[\delta \dot{x} = A \delta x + B \delta u + E \delta d, \qquad \delta y = C \delta x + D \delta u + F \delta d,\]

where the matrices are Jacobians evaluated at the operating point.

What this means physically is that even nonlinear plants often look locally linear near the regime where we intend to control them. Classical control lives inside that local picture.


4. Dynamics Before Control

Before drawing a feedback loop, we should ask how the plant behaves by itself.

4.1 Natural response and forced response

Any linear response can be understood as the sum of:

For the heated chamber with $u=d_\theta=0$,

\[\theta(t) = \theta(0)e^{-t/\tau}.\]

For an underdamped second-order system,

\[x_n(t) = e^{-\zeta \omega_n t}\left(A \cos \omega_d t + B \sin \omega_d t\right),\]

where

\[\omega_d = \omega_n \sqrt{1-\zeta^2}.\]

What this means physically is that the unforced system already contains the seeds of its future behavior. Control does not create dynamics from nothing; it reshapes the existing ones.

4.2 Equilibrium and operating behavior

An equilibrium is a condition where the state does not change if the inputs remain fixed.

For the thermal plant, equilibrium means heating exactly balances heat loss. For the mechanical plant, equilibrium means force balance with zero velocity and zero acceleration.

Once a plant is perturbed away from equilibrium, the most important questions are:

Those are the raw ingredients of control performance.

4.3 Transient versus steady behavior

Control design always balances two time scales:

Students often jump too fast to steady-state error formulas. But a real controller is judged by both:

4.4 Poles and eigenvalues as geometry of motion

For linear systems, poles or eigenvalues are the compact summary of natural motion.

What this means physically is that pole location is a geometry of speed and oscillation. Control design will eventually become the art of moving that geometry.


5. What Stability Analysis Is Trying to Answer

Stability is not a decorative theorem topic. It is the first pass/fail test of whether a controller deserves to exist.

5.1 Internal stability and BIBO stability

Two stability ideas are worth separating.

Internal stability asks:

If I perturb the state and then stop exciting the system, do the internal variables return to equilibrium?

BIBO stability asks:

If the input stays bounded, does the output stay bounded?

For minimal LTI systems these notions line up closely, but conceptually they answer different questions. Internal stability cares about hidden modes. BIBO stability cares about what the external world sees at the output.

5.2 Asymptotic, marginal, and unstable behavior

For continuous-time LTI systems:

Examples:

5.3 Why stability comes before tuning

The question is not “Can I reduce the error?” The question is “Can I reduce the error without creating divergence or unacceptable oscillation?”

This is why classical design tools are organized around stability:

Only after stability is understood does controller tuning become meaningful.

5.4 Stability is necessary, not sufficient

A stable loop may still be bad:

So control design is never only “stabilize the plant.” It is “stabilize it with acceptable performance and robustness.”


6. LTI Modelling Tools: Laplace, Transfer Functions, and Pole-Zero Maps

Once the model is linear and time-invariant around an operating point, Laplace transforms turn differential equations into algebra.

6.1 Transfer function under zero initial conditions

The transfer function is defined by

\[G(s) = \frac{Y(s)}{U(s)}\]

under zero initial conditions.

That zero-initial-condition phrase matters. A transfer function describes how input shapes output; it does not by itself carry the full memory of arbitrary initial state.

6.2 Transfer functions of the running examples

For the heated chamber,

\[G_\theta(s) = \frac{\Theta(s)}{U(s)} = \frac{K}{\tau s + 1}.\]

For the mass-spring-damper,

\[G_x(s) = \frac{X(s)}{U(s)} = \frac{1}{m s^2 + b s + k}.\]

What this means physically is that the first plant has one dominant mode and the second has two interacting modes. The order of the denominator reflects the amount of dynamic memory that matters.

6.3 Poles and zeros

Poles govern natural motion. Zeros shape how inputs are filtered and how transient responses are bent.

Zeros do not by themselves create instability, but they can make control harder, especially when they introduce phase lag or non-minimum-phase behavior.

6.4 Why transfer functions are so useful

Transfer functions are powerful because they compress several ideas into one object:

This is why classical control can move quickly between equations, geometry, and design plots.


7. Feedback Structure and Closed-Loop Logic

Now that the plant model exists, control can begin.

7.1 Open loop versus closed loop

In open-loop control, we choose the actuator input without checking whether the output followed the command.

In closed-loop control, we measure the output, compare it with the reference, and drive the error toward zero.

The common negative-feedback structure is:

Closed loop is powerful because it reacts to disturbances and model mismatch rather than hoping the original command was perfect.

7.2 Closed-loop transfer function

For the standard loop,

\[E(s) = R(s) - H(s)Y(s), \qquad U(s) = D(s)E(s), \qquad Y(s) = G(s)U(s).\]

Hence,

\[\frac{Y(s)}{R(s)} = \frac{D(s)G(s)}{1 + D(s)G(s)H(s)}.\]

For unity feedback, $H(s)=1$ and the loop gain is

\[L(s) = D(s)G(s).\]

What this means physically is that control changes the denominator of the system. The closed-loop characteristic equation

\[1 + L(s) = 0\]

determines the new poles, so feedback is literally a way of reshaping the plant’s dynamics.

7.3 Disturbance rejection and measurement tradeoffs

Feedback is attractive because the same loop that tracks reference changes can also reject disturbances. But the loop cannot improve everything at once:

So the feedback loop is a tradeoff machine, not a free lunch.


8. Time-Domain Analysis: Speed, Damping, and Accuracy

Once the loop is closed, we ask what the transient and steady response look like.

8.1 Step-response specifications

The usual time-domain performance measures are:

These are not arbitrary metrics. They are compact proxies for speed, damping, and accuracy.

8.2 Standard second-order template

The most important reference model is

\[G_{cl}(s) = \frac{\omega_n^2}{s^2 + 2 \zeta \omega_n s + \omega_n^2}.\]

Here:

Useful memory formulas are

\[M_p \approx e^{-\pi \zeta/\sqrt{1-\zeta^2}} \times 100\%,\] \[t_s \approx \frac{4}{\zeta \omega_n}.\]

What this means physically is:

8.3 Final value and steady-state error

When the closed loop is stable, the final-value theorem gives

\[\lim_{t \to \infty} y(t) = \lim_{s \to 0} sY(s).\]

This lets us reason about steady-state error for step, ramp, and disturbance inputs.

One particularly important classical idea is system type, meaning the number of integrators in the open loop:

That is why integral action is so common in servo and process control, and also why it must be added carefully.


9. Classical Stability Tests and Design Views

Once we know what kind of performance we want, classical control gives several complementary views of stability and design.

9.1 Routh-Hurwitz criterion

Routh-Hurwitz tests whether a characteristic polynomial has any right-half-plane roots without explicitly solving for all of them.

For the cubic

\[s^3 + a_2 s^2 + a_1 s + a_0,\]

stability requires

\[a_2 > 0, \qquad a_1 > 0, \qquad a_0 > 0, \qquad a_2 a_1 > a_0.\]

What this means physically is that we can rule out unstable closed-loop pole patterns directly from the controller-dependent polynomial. It is one of the fastest algebraic sanity checks in classical design.

9.2 Root locus

Root locus answers the question:

If I change a gain parameter, where do the closed-loop poles go?

The key picture is:

This is why root locus is especially useful for first passes with P, PI, lead, and lag design.

9.3 Pole-zero plots

Pole-zero plots are the static geometric summary. Root locus adds the dynamic “what happens when the gain changes?” story.

Together they teach an important idea: controller tuning is pole shaping, whether done graphically, algebraically, or with state-space methods.

9.4 Bode plots

Bode plots evaluate the open-loop or closed-loop response across frequency.

For a first-order low-pass element,

\[G(s) = \frac{1}{\tau s + 1},\]

we get

\[|G(j\omega)| = \frac{1}{\sqrt{1+(\omega \tau)^2}}, \qquad \angle G(j\omega) = -\tan^{-1}(\omega \tau).\]

What this means physically is that the plant follows slow variations well but lags and attenuates rapid variations. The controller must work with that limitation, not wish it away.

9.5 Gain margin and phase margin

Margins quantify how much extra uncertainty the loop can tolerate before it loses stability.

These are not just textbook numbers. They are practical robustness indicators against unmodelled dynamics, delay, and sensor/actuator imperfections.

9.6 Nyquist and Nichols

Nyquist connects the open-loop frequency plot to exact closed-loop stability:

\[N = Z - P,\]

where $N$ counts encirclements of $-1$, $P$ is the number of open-loop right-half-plane poles, and $Z$ is the number of closed-loop right-half-plane poles.

Nichols charts compress gain and phase into one plane with closed-loop performance contours. They are useful when thinking directly in loop-shaping terms rather than alternating between separate magnitude and phase plots.

Together, Bode, Nyquist, and Nichols are three views of the same problem: frequency-domain stability and robustness.


10. Sensitivity, Complementary Sensitivity, and Loop Shaping

Classical control becomes more coherent once everything is expressed through the loop gain

\[L(s) = D(s)G(s)H(s).\]

10.1 Sensitivity functions

Two central closed-loop quantities are

\[S(s) = \frac{1}{1+L(s)}, \qquad T(s) = \frac{L(s)}{1+L(s)}.\]

Here:

What this means physically is:

Because

\[S(s) + T(s) = 1,\]

the controller cannot make both small everywhere. That is the core loop-shaping tradeoff.

10.2 Practical loop-shaping goals

In many servo and process loops, we want:

So “good control” rarely means “large gain.” It means gain placed in the right frequency regions.

10.3 Why loop shaping unifies the classical toolbox

Root locus, Bode, Nyquist, Nichols, margins, and compensator design can feel like unrelated chapters. Loop shaping is what unifies them.

Each tool is answering the same question in a different language:

How should we reshape the open-loop dynamics so that the closed loop becomes fast enough, accurate enough, and robust enough?


11. Controllers as Model-Shaping Tools

Controllers are easiest to understand once we stop treating them as mysterious formulas and instead view them as deliberate ways of changing low-frequency gain, crossover behavior, damping, and steady-state accuracy.

11.1 P, PI, and PID

The standard PID controller is

\[D(s) = K_P + \frac{K_I}{s} + K_D s.\]

Roles of the terms:

What this means physically is:

Most real loops end up at PI or PID because pure proportional control is often too biased in steady state, while integral action alone is too blunt without damping support.

11.2 Lead and lag compensation

A lead compensator has the form

\[D_{lead}(s) = K \frac{1+\tau s}{1+\alpha \tau s}, \qquad 0 < \alpha < 1.\]

A lag compensator has the form

\[D_{lag}(s) = K \frac{1+\tau s}{1+\beta \tau s}, \qquad \beta > 1.\]

Interpretation:

11.3 Anti-windup

Actuators saturate. When they do, the integral term may continue accumulating error that the actuator cannot physically realize. That is integrator windup.

Anti-windup schemes keep the controller’s internal memory consistent with actuator reality, so recovery after saturation is not unnecessarily slow or oscillatory.

11.4 Servo design, feedforward, and 2-DOF thinking

In reference-tracking problems, feedback alone is not always the cleanest structure.

Useful additions are:

This is often where classical design starts to feel like engineering rather than pure mathematics.


12. The State-Space Bridge: Internal Structure, Controllability, and Observability

Transfer functions are excellent for single-loop classical design, but they hide internal structure.

12.1 Why transfer functions become limiting

Transfer functions answer an input-output question. They are less natural when:

That is why modern control reorganizes the same problem in state-space form.

12.2 State-space model

The standard linear model is

\[\dot{x} = Ax + Bu, \qquad y = Cx + Du.\]

The eigenvalues of $A$ are the internal-mode version of the pole picture. For minimal realizations, they line up with the transfer-function poles.

12.3 Controllability and observability

The controllability matrix is

\[\mathcal{C} = \begin{bmatrix} B & AB & \cdots & A^{n-1}B \end{bmatrix}.\]

The observability matrix is

\[\mathcal{O} = \begin{bmatrix} C \\ CA \\ \vdots \\ CA^{n-1} \end{bmatrix}.\]

Full rank means:

What this means physically is that no clever controller can move a mode it cannot actuate, and no clever observer can reconstruct a mode that never shows up in measured data.

12.4 Full-state feedback and pole placement

With state feedback,

\[u = -Kx + Nr,\]

the closed-loop matrix becomes

\[A_{cl} = A - BK.\]

This is the modern form of the same idea classical control has been using all along: move the poles to reshape the dynamics.

For the full state-space, observer, MIMO, and robust multivariable story, continue to Modern, Multivariable, and Networked Control.


13. The Bridge to Nonlinear Control

Classical control mostly lives in a locally linear world, but physical plants are rarely perfectly linear.

13.1 Phase plane and local behavior

For second-order systems, the phase plane gives a geometric picture of state trajectories:

This is the simplest doorway into nonlinear thinking because it makes stability visible as geometry rather than just algebra.

13.2 Linear models are local, nonlinear behavior is global

Linearization near an operating point is often extremely useful, but only locally.

Outside that region, real plants may show:

That is why many practical controllers are hybrids:

For the broader nonlinear story, including Lyapunov methods, backstepping, and sliding mode, continue to Topics in Nonlinear Systems.


14. Reading Map Across the Control Series

Use this note for the first pass through control logic:

  1. what the plant is,
  2. how the model is built,
  3. how dynamics and stability are interpreted,
  4. how feedback reshapes the poles,
  5. how classical tools express performance and robustness.

Then branch outward:


15. Compact Recall Map

When control theory starts to feel like too many disconnected formulas, compress it to this logic:

  1. A controller acts on a system, not on an equation in isolation.
  2. A useful model identifies states, inputs, outputs, disturbances, and operating assumptions.
  3. The plant has its own dynamics before any controller is added.
  4. Stability analysis asks whether those dynamics return, drift, or diverge.
  5. Feedback changes the characteristic equation and therefore reshapes the poles.
  6. Time-domain tools describe speed, damping, and accuracy.
  7. Frequency-domain tools describe robustness, crossover, margins, and noise tradeoffs.
  8. Compensators such as PID, lead, and lag are ways of shaping the loop, not isolated formulas.
  9. State-space methods reveal which internal modes can be controlled and observed.
  10. Nonlinear methods take over when local linear reasoning is no longer enough.

That is the natural logic of basic control: model the system, understand the dynamics, analyze stability, then design feedback to reshape behavior.