Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

Introduction and Motivation for PDE-Constrained Optimal Control

University of Pisa

Overview

This lecture introduces optimal control as an optimization problem constrained by equations. We first build the finite-dimensional analogy, then move to PDE-constrained models.

Logical path of the lecture:

  1. start from a constrained optimization problem in (y,u)(y,u);

  2. eliminate the state through the model equation when possible;

  3. obtain a reduced optimization problem in the control variable only;

  4. derive first-order conditions in finite dimension;

  5. transfer the same structure to PDE settings.


General Optimal Control Problem

Choose a control variable uu and a state variable yy such that

(y,u)argmin(y,u)J(y,u)(y^\star,u^\star)\in\operatorname*{argmin}_{(y,u)} J(y,u)

subject to

E(y,u)=0,uUad.\mathcal{E}(y,u)=0,\qquad u\in\mathcal{U}_{\mathrm{ad}}.

Key ingredients:


Forward Problem vs Control Problem

Forward problem:

Optimal control:


Finite-Dimensional Setting (Simultaneous vs Reduced)

Consider

minyRn,uRmJ(y,u)s.t.Ay=Bu,\min_{y\in\mathbb{R}^n,\,u\in\mathbb{R}^m} J(y,u) \quad\text{s.t.}\quad Ay=Bu,

with ARn×nA\in\mathbb{R}^{n\times n} invertible.

Simultaneous formulation

Optimize in (y,u)(y,u) and enforce Ay=BuAy=Bu explicitly.

Interpretation:

Reduced formulation

Since AA is invertible,

y=A1Bu=:S(u),y=A^{-1}Bu=:S(u),

where SS is the control-to-state operator. Then define

f(u):=J(S(u),u),f(u):=J(S(u),u),

and solve

minuUadf(u).\min_{u\in\mathcal{U}_{\mathrm{ad}}} f(u).

This reduces a finite-dimensional control problem to a standard finite-dimensional optimization problem.

Step-by-step logic:

  1. the constraint Ay=BuAy=Bu defines yy uniquely as a function of uu;

  2. therefore the only independent decision variable is uu;

  3. the objective becomes a composite map uJ(S(u),u)u\mapsto J(S(u),u);

  4. all constraint information is encoded in SS.


Existence in Finite Dimensions

A standard existence result for the reduced problem:

If

then a minimizer exists.

Reason: in finite dimensions, closed and bounded sets are compact (Weierstrass theorem).

Why this matters for control:


Unconstrained First/Second-Order Conditions

For convex differentiable ff on a convex set KK:

f(uˉ)(uuˉ)0uK\nabla f(\bar u)\cdot (u-\bar u)\ge 0\quad\forall u\in K

is the first-order optimality condition.

Special case (interior point):

f(uˉ)=0.\nabla f(\bar u)=0.

If fC2f\in C^2 and uˉ\bar u is a local minimizer:

If moreover D2f(uˉ)D^2 f(\bar u) is positive definite, then uˉ\bar u is a strict local minimizer.


Constrained minimization (2D Example)

Let

Uad=R2,f(u)=12uTAu,φ(u)=Bug=0,\mathcal{U}_{\mathrm{ad}}=\mathbb{R}^2,\qquad f(u)=\frac12 u^T A u,\qquad \varphi(u)=Bu-g=0,

with

AR2×2 SPD,BR1×2,gR.A\in\mathbb{R}^{2\times 2}\text{ SPD},\quad B\in\mathbb{R}^{1\times 2},\quad g\in\mathbb{R}.

Logical interpretation:

  1. objective level sets are ellipses (because AA is SPD);

  2. feasible points lie on an affine line Bug=0Bu-g=0;

  3. the minimizer is where the first objective level set touches that feasible line.

At an optimal feasible point, f\nabla f is orthogonal to the feasible tangent direction, so it must be parallel to φ\nabla\varphi:

f(uˉ)=(φ(uˉ))Tλ.\nabla f(\bar u)=(\nabla\varphi(\bar u))^T\lambda.

Equivalent first-order form:

f(uˉ)(φ(uˉ))Tλ=0,φ(uˉ)=0.\nabla f(\bar u)-(\nabla\varphi(\bar u))^T\lambda=0, \qquad \varphi(\bar u)=0.

Meaning:

Lagrangian Formalism

Define the Lagrangian

L(u,λ)=f(u)φ(u)λ.\mathcal{L}(u,\lambda)=f(u)-\varphi(u)\cdot\lambda.

and search for saddle points:

uˉ,λˉ=argminuargmaxλL(u,λ).\bar u, \bar\lambda = \arg\min_u \arg\max_\lambda \mathcal{L}(u,\lambda).

Stationarity gives

Lu=0,Lλ=0.\frac{\partial\mathcal{L}}{\partial u}=0, \qquad \frac{\partial\mathcal{L}}{\partial \lambda}=0.

For this quadratic/affine case:

uL=AuBTλ=0,λL=Bu+g=0,\nabla_u\mathcal{L}=Au-B^T\lambda=0, \qquad \nabla_\lambda\mathcal{L}=-Bu+g=0,

which is the KKT linear system

(ABT B0)(u λ)=(0 g).\begin{pmatrix}A & -B^T \\\ -B & 0\end{pmatrix}\begin{pmatrix}u \\\ \lambda\end{pmatrix}=\begin{pmatrix}0 \\\ -g\end{pmatrix}.

This is the prototype for all later optimality systems:

Geometry of the 2D minimization example Objective restricted to the feasible line

From Finite to Infinite Dimensions

For PDE-constrained control, state/control live in function spaces (typically Hilbert spaces):

Typical difficulties:

Conceptual continuity with the finite-dimensional case:

  1. same optimization structure;

  2. same reduced-vs-simultaneous viewpoints;

  3. same KKT logic;

  4. only the functional-analytic setting changes.


Prototypical PDE-Constrained Models

Elliptic distributed control

min(y,u)12yydL2(Ω)2+α2uL2(Ω)2\min_{(y,u)} \frac12\|y-y_d\|_{L^2(\Omega)}^2 +\frac\alpha2\|u\|_{L^2(\Omega)}^2

subject to

Δy=u in Ω,y=0 on Ω,uUad.-\Delta y=u\text{ in }\Omega, \qquad y=0\text{ on }\partial\Omega, \qquad u\in U_{\mathrm{ad}}.

Parabolic control

tyΔy=u in Q,y(,0)=y0,\partial_t y-\Delta y=u\text{ in }Q, \qquad y(\cdot,0)=y_0,

with tracking over space-time and Tikhonov regularization.

Flow and inverse problems


Control Constraints

Common box constraints:

uminuumax.u_{\min}\le u\le u_{\max}.

They lead to variational inequalities and KKT systems in function spaces.


Typical Variations of OCP Formulations

Many optimal control models keep the same abstract structure but vary in where the control acts, what is observed, and how the objective is measured.

Common variations include:

These variants motivate why we need a flexible theoretical framework and multiple numerical methods in the rest of the course.


References for This Lecture