Model predictive control (MPC) has demonstrated exceptional success for the high-performance control of complex systems [1], [2]. The conceptual simplicity of MPC as well as its ability to effectively cope with the complex dynamics of systems with multiple inputs and outputs, input and state/output constraints, and conflicting control objectives have made it an attractive multivariable constrained control approach [1]. MPC (a.k.a. receding-horizon control) solves an open-loop constrained optimal control problem (OCP) repeatedly in a receding-horizon manner [3]. The OCP is solved over a finite sequence of control actions {u0,u1,f,uN- 1} at every sampling time instant that the current state of the system is measured. The first element of the sequence of optimal control actions is applied to the system, and the computations are then repeated at the next sampling time. Thus, MPC replaces a feedback control law p(m), which can have formidable offline computation, with the repeated solution of an open-loop OCP [2]. In fact, repeated solution of the OCP confers an implicit feedback action to MPC to cope with system uncertainties and disturbances. Alternatively, MPC approaches circumvent the need to solve an OCP online by deriving relationships for the optimal control actions in terms of an explicit function of the state and reference vectors. However, MPC is not typically intended to replace standard MPC but, rather, to extend its area of application [4]-[6].