Professional Documents
Culture Documents
Predictive Control
An Overview and Perspectives
for Future Research
Ali Mesbah
M
odel predictive control (MPC) has demon- peated solution of the OCP
strated exceptional success for the high-per- confers an “implicit” feed-
formance control of complex systems [1], [2]. back action to MPC to cope
The conceptual simplicity of MPC as well as with system uncertainties
its ability to effectively cope with the complex and disturbances. Alterna-
dynamics of systems with multiple inputs and outputs, in- tively, explicit MPC approaches image licensed by ingram publishing
put and state/output constraints, and conflicting control circumvent the need to solve an
objectives have made it an attractive multivariable con- OCP online by deriving relationships
strained control approach [1]. MPC (a.k.a. receding-horizon for the optimal control actions in terms of an “explicit”
control) solves an open-loop constrained optimal control function of the state and reference vectors. However, explic-
problem (OCP) repeatedly in a receding-horizon manner it MPC is not typically intended to replace standard MPC
[3]. The OCP is solved over a finite sequence of control ac- but, rather, to extend its area of application [4]–[6].
tions {u 0, u 1, f, u N - 1} at every sampling time instant that the Although MPC offers a certain degree of robustness to
current state of the system is measured. The first element system uncertainties due to its receding-horizon imple-
of the sequence of optimal control actions is applied to the mentation, its deterministic formulation typically renders
system, and the computations are then repeated at the next it inherently inadequate for systematically dealing with
sampling time. Thus, MPC replaces a feedback control law uncertainties. Addressing the general OCP for uncertain sys-
r ($), which can have formidable offline computation, with tems would involve solving the dynamic programming (DP)
the repeated solution of an open-loop OCP [2]. In fact, re- problem over (arbitrary) feedback control laws r ($) [7], [8].
Solving the DP problem, however, is deemed to be imprac-
Digital Object Identifier 10.1109/MCS.2016.2602087 tical for real-world control applications since computational
Date of publication: 11 November 2016 complexity of a DP problem increases exponentially with
generation and distribution, chemical processes, operation Prx [·] : = Pr [·|x (0) = x] is the conditional probability.
research, networked controlled systems, and vehicle path x A : = x < Ax is the weighted two-norm of x, where A is a
planning. Table 1 provides an overview of various emerg- positive-definite matrix.
ing application areas for SMPC; this table by no means pro-
vides an exhaustive list of SMPC applications reported in General Formulation of SMPC
the literature. The majority of reported SMPC approaches Consider a stochastic, discrete-time system
have been developed for linear systems (for example, algo-
rithms based on the stochastic tube [26] or affine parame- x t + 1 = f (x t, u t, w t), (1a)
terization of control policy [27]). Several SMPC applications
y t = h (x t, u t, v t), (1b)
to linear and nonlinear systems have been reported based
on stochastic programming-based approaches [28]–[30]
and Monte Carlo sampling techniques [31], [32]. Stochastic where t ! N + ; x t ! R n x, u t ! U 1 R n u , and y t ! R n y are the
nonlinear MPC (SNMPC) has received relatively little system states, inputs, and outputs, respectively; U is a non-
attention, with only a few applications reported mainly in empty measurable set for the inputs; w t ! R n w and v t ! R n v
the area of process control [33]–[35]. are disturbances and measurement noise that are unknown
This article gives an overview of the main developments at the current and future time instants but have known
in the area of SMPC in the past decade and provides the probability distributions pw and pv , respectively; and f and
reader with an impression of the different SMPC algorithms h are (possibly nonlinear) Borel-measurable functions that
and the key theoretical challenges in stochastic predictive describe the system dynamics and outputs, respectively.
control without undue mathematical complexity. The gen- For simplicity, the formulation of SMPC is presented for
eral formulation of a stochastic OCP is first presented, fol- the case of full state-feedback control, in which the system
lowed by an overview of SMPC approaches for linear and states are known at each sampling time instant. Let N ! N
nonlinear systems. Suggestions of some avenues for future be the prediction horizon, and assume that the control hori-
research in this rapidly evolving field concludes the article. zon is equal to the prediction horizon. Define an N -stage
feedback control policy as
Notation
R n denotes the n-dimensional Euclidean space. R + : = [0, 3) . r := {r 0 ($), r 1 ($), f, r N - 1 ($)}, (2)
N = {1, 2, f} is the set of natural numbers. N + : = N , {0} .
Z [a, b] : = {a, a + 1, f, b} is the set of integers from a to b. px is where the Borel-measurable function r i ($) : R (i + 1) n x " U, for
the (multivariate) probability distribution of random all i = 0, f, N - 1 is a general (causal) state feedback con-
variable(s) x. E denotes expectation and E x [·] : = E [·|x (0) = x] trol law. At the ith stage of the control policy, the control
is the conditional expectation. Pr denotes probability, and input u i is selected as the feedback control law r i ($), that is,
Linear Nonlinear
System dynamics [42]–[59], [61], [62], [69], [73], [89], [91], [92], [147] [31]–[35], [63]–[65], [110], [120]
Type of uncertainties Time-varying uncertainties Time invariant Time-varying/invariant
uncertainties multiplicative
Additive Multiplicative
[33], [34], [65] uncertainties with
[43]–[45], [47]–[53], [55], [57], [58], [69]
additive disturbances
[56], [61], [73], [92], [147]
[26], [42], [54], [59],
[62], [89]
Uncertainty propagation Stochastic tube Scenario/sample based gPC/FP/GM
[26], [42]–[46] [31], [32], [56], [33]–[35], [54], [64], [65], [89]
[59]–[63], [92], [93]
Control input Open-loop control Prestabilizing feedback Affine parameterization of control policy
parameterization actions control [46]–[56]
[31]–[35], [57]–[59], [26], [42]–[44]
[61]–[65], [89], [92],
[93], [120]
Input constraint Hard input constraints Probabilistic input Saturated input
[33]–[35], [47], [65] constraints constraints
[48], [69], [73] [50], [52], [55]
Chance constraint Individual Joint Expectation type
[26], [33], [42]–[45], [47], [58], [110] [34], [35], [55], [56], [52], [62], [65], [69]
[61], [73], [89], [93]
Receding-horizon Full-state feedback Output feedback
implementation (The majority of the reported work) [45], [52], [73], [77]
Optimization algorithm Convex QP/SOCP SDP Stochastic programing Nonconvex
[43], [44], [47], [50]–[53], [69] [57], [58], [91] optimization
[55], [89], [147] [31]–[35], [63]–[65],
[120]
q
[70], the states x t are expressed in terms of the sum of a
x t + 1 = Ax t + Bu t + / (Ar j x t + Br j u t) w t,j, (8) deterministic component z t and a random component e t
j= 1
q
where {w t, j} j = 1 is a sequence of zero-mean i.i.d. random x t := z t + e t, z t + 1 = Uz t + Bo t, e t + 1 = Ue t + Dw t,
variables with known variance (see, for example, [68] and
[69]). An overview of the main SMPC approaches for linear where u t := Kx t + o t in (7a) with {o 0, f, o N - 1} is the decision
systems is given below. variables in the stochastic OCP and K is a constant feedback
gain, and U := A + BK is assumed to be strictly stable. When
Stochastic Tube Approaches the disturbances w t are Gaussian, the representation
Stochastic tube approaches to MPC of linear systems with x t := z t + e t allows for replacing linear chance constraints of
additive, bounded disturbances are presented in [43] and the form Pr [cx i # d] $ b with tightened hard constraints
[44] with the objective of minimizing the infinite-horizon cx i # dl # d , so that cz i # dl implies Pr [cx i # d] $ b, for all
value function i = 0, f, N - 1 [43]. Hence, stochastic tubes enable replac-
ing the state chance constraints with linear constraints
defined in terms of the nominal predictions z i . This leads to
V3 (x t, r) = E x t ; / xt i + u i 2RE , (9)
3
2
Q
i=0 a significant reduction in the number of decision variables
in the online optimization. Tightened constraints can also
subject to state chance constraints. Stochastic tubes are be determined (with more complex offline computations)
deployed to provide guarantees for recursive feasibility when w t has an arbitrary, but known, distribution [44].
and, as a result, to ensure closed-loop stability and con- To obtain a computationally efficient algorithm, sto-
straint satisfaction. Like in tube-based MPC [12], [13], [15], chastic tubes with fixed ellipsoidal cross sections are used
per stage and the final cost functions are defined in terms computationally tractable algorithm, the gPC framework
of the expected values of future nominal states and the [85], [86] with Galerkin projection [87] is used for propagat-
variances of future errors in the estimated states. The con- ing the time-invariant uncertainties through the system
trol actions {o 0, f, o N - 1} and the feedback gains dynamics (the gPC framework was first adopted in [88] to
{K 0, f, K N - 1} are used to drive the mean of the states to address the problem of linear-quadratic regulator design).
zero and minimize the variance of states, respectively. Ter- An affine-state parameterization is used to minimize the
minal constraints are included in the stochastic OCP to expectation of a quadratic cost-per-stage function over a
guarantee recursive feasibility [3]. To handle unbounded finite prediction horizon. Under the assumptions of per-
disturbances that can lead to infeasibility of the problem, fect state information and stabilizability of the pair (A, B)
an initialization constraint is included in the OCP to allow for all uncertainty realizations, the closed-loop stability of
the optimizer to select the initial conditions based on the the SMPC algorithm is established for the unconstrained
feasibility and optimality requirements of the control algo- case by bounding the value function. For a linear system
rithm. The main shortcomings of the SMPC algorithm pre- with a similar setup, a polynomial chaos-based SMPC
sented in [73] are 1) an inability to consider saturation algorithm is also reported in [89]. This SMPC algorithm,
functions in the control policy to enable handling hard however, solves the stochastic OCP merely over the control
input bounds (as in [52]), 2) the conservatism associated actions y (without considering any feedback control
with the Chebyshev–Cantelli inequality used for chance parameterization) and does not provide guarantees on
constraint approximation, and 3) nonconvexity of the algo- closed-loop stability. An SMPC algorithm for linear sys-
rithm. Using a problem setup similar to that in [52] and [73], tems where the matrices A and B consist of multiplicative,
an SMPC approach is presented in [55] that can handle joint unbounded disturbances [see (8)] is presented in [69]. The
state chance constraints and hard input constraints under algorithm in [69] considers expectation-type constraints on
closed-loop prediction in the presence of arbitrary (possi- inputs and states and establishes the closed-loop stability
bly unbounded) additive disturbances. The SMPC approach and constraint satisfaction properties of the control algo-
uses risk allocation [83], [84] in combination with the Can- rithm by defining a terminal constraint. This algorithm
telli–Chebyshev inequality to obtain computationally trac- uses the feedback control law u t = K t x t + o t to transform
table surrogates for the joint state-chance constraints when the stochastic OCP into a semidefinite programming (SDP)
the first two moments of the arbitrary disturbance distribu- problem, in which the feedback gains k t and control
tions are known. An algorithm is then presented for solv- actions y t are computed online. The SDP formulation can
ing two convex optimization problems iteratively to render the SMPC algorithm computationally involved for
determine the best feedback gain and optimal risk alloca- high-dimensional systems.
tion. Simulation studies show that optimizing the risk allo- To summarize, affine-disturbance and affine-state
cation in [55] can lead to improved control performance as parameterizations of a feedback control policy have been
compared to an MPC algorithm with fixed risk allocation. widely used to obtain convex SMPC algorithms. In the pres-
Reference [54] addresses the problem of predictive con- ence of unbounded disturbances, dealing with hard input
trol of linear systems with arbitrary time-invariant proba- bounds and establishing closed-loop stability pose key
bilistic uncertainties and additive Gaussian disturbances challenges for these algorithms. An effective approach to
address these challenges is to introduce a (nonlinear) satu-
x t + 1 = A (d) x t + B (d) u t + Dw t, (10) ration function to the affine parameterization of a feedback
control policy. The saturated feedback control laws not only
with d denoting the time-invariant uncertain system enable handling of hard input bounds for systems with
parameters with known (finite-variance) probability dis- unbounded disturbances but also facilitate ensuring mean-
tributions and w t being zero-mean Gaussian disturbance. square boundedness of states when the system is Lyapunov
In contrast to the above SMPC approaches (see, for exam- stable. Closed-loop stability of these SMPC algorithms is
ple, [47] and [52]), the algorithm in [54] handles system commonly guaranteed by defining a negative drift condi-
descriptions in which the system matrices are nonlinear tion via either a stability constraint or appropriate selection
ly dependent on stochastic uncertainties. To obtain a of the value function. Affine control policies have also been
scenario fans, which consist of bunches of independent and sequential conic-programming algorithm that is computa-
equiprobable scenarios. However, scenario trees, in which sce- tionally expensive for high-dimensional systems with
narios are not equiprobable, provide a causal representation many state constraints and long prediction horizons. To
of the uncertainties acting on the system, leading to more tackle the computational complexity of the algorithm, the
compact scenario representations [105]. problem is decomposed using the linear-quadratic Gaussian
A technique for fast scenario removal based on mixed- decomposition, where the feedback and feedforward control
integer quadratic optimization is proposed in [61]. On the problems are solved separately.
other hand, a scenario-based SMPC algorithm is intro- In [31] and [63], the use of a Markov-chain Monte Carlo
duced in [62] in which chance constraints are replaced with (MCMC) technique [111] is proposed for solving constrained
time-averaged, instead of pointwise-in-time, constraints to nonlinear stochastic optimization problems. A well-estab-
reduce the theoretically required number of scenarios. It is lished theoretical framework (including general convergence
shown in [62] that the sample size is independent of the results and central limit theorems) exists for MCMC approach-
state dimension and, in fact, the sample size is dictated by es as well as sequential Monte Carlo approaches under weak
the support rank of state constraints. This implies that a assumptions (see, for example, [112]). The algorithms in [31]
small sample size would be sufficient even for a system and [63] do not rely on convexity assumptions. In addition,
with high state dimension as long as the support rank of they guarantee convergence (in probability) to a near-optimal
state constraints is low. The algorithm in [62] also allows for solution under mild conditions related to the convergence of a
handling multiple state-chance constraints (see [106]). homogeneous Markov chain. However, the effectiveness of
Despite the extensive work done in the area of scenario- these stochastic optimal control algorithms has not been dem-
based MPC, the primary challenge still lies in identifying the onstrated when implemented in a receding-horizon manner. A
appropriate number of scenarios that not only guarantees an sequential Monte Carlo technique is adopted in [32] to reduce
admissible level of constraint satisfaction but also makes the the computational complexity associated with MCMC ap-
computational requirements of the algorithm manageable for proaches in [31] and [63].
practical control applications. Hence, recent work on sce- A key challenge in SMPC of nonlinear systems is the
nario-based MPC has mainly attempted to reduce the com- efficient propagation of stochastic uncertainties through
putational complexity of these algorithms [107]–[109]. Further the system dynamics. For discrete-time nonlinear systems
development in this area is crucial to facilitate more practical with additive disturbances, the Gaussian-mixture approxi-
applications of scenario-based control approaches. Another mation [113] is used in [64] to describe the transition prob-
challenging problem in scenario-based MPC arises from ability distributions of states in terms of weighted sums of
establishing the theoretical properties such as recursive feasi- a predetermined number of Gaussian distributions. Due to
bility and closed-loop stability, in particular for the case of its universal approximation property, the method of Gauss-
unbounded uncertainties. ian-mixture approximation provides a flexible framework
for constructing the probability distributions of stochastic
Stochastic Model Predictive variables and, therefore, for uncertainty propagation. The
Control for Nonlinear Systems knowledge of complete probability distributions allows for
MPC of stochastic nonlinear systems has received relatively defining the value function of the stochastic OCP in terms
little attention. A pioneering work in this area is the SNMPC of the complete distributions of stochastic states, instead of
approach presented in [110]. The algorithm relies on the merely some moments of the probability distributions.
notion of optimizing a deterministic feedforward trajectory However, online adaptation of the weights of the Gaussian
for constraint handling and optimizing a linear time-varying distributions, which is essential for describing the time
feedback controller for minimizing the closed-loop variance evolution of the probability distributions, poses a computa-
around the reference trajectory (u t = K t x t + o t ). To explicitly tional challenge in Gaussian-mixture approximations.
account for back-off with respect to the reference trajectory, The gPC framework [85], [86] is used in [33] to develop an
the nonlinear system dynamics are linearized around the ref- SNMPC approach for a general class of nonlinear systems
erence trajectory. The control algorithm is rendered convex that are subject to arbitrary time-invariant stochastic uncer-
through the Youla–Kuc˘era parameterization, and the chance tainties in system parameters and initial conditions. The
constraints are reduced to second-order cone constraints gPC framework replaces the implicit mappings between
using an ellipsoidal relaxation. The cone constraints lead to a uncertain variables/parameters and states (defined in terms