Abstract

In this paper, we compute a (local) optimal reduced order model that matches a prescribed set of moments of a stable linear time-invariant system of high dimension. We fix the interpolation points and parametrize the models achieving moment-matching in a set of free parameters. Based on the parametrization and using the H2-norm of the approximation error as the objective function, we derive a nonconvex optimization problem, i.e., we search for the optimal free parameters to determine the model yielding the minimal H2-norm of the approximation error. Furthermore, we provide the necessary first-order optimality conditions in terms of the controllability and the observability Gramians of a minimal realization of the error system. We then propose two gradient-type algorithms to compute the (local) optimal models, with mathematical guarantees on the convergence. We also derive convex semidefinite programming relaxations for the nonconvex Problem, under the assumption that the error system admits block-diagonal Gramians, and derive sufficient conditions to guarantee the block diagonalization. The solutions resulting at each step of the proposed algorithms guarantee the achievement of the imposed moment matching conditions. The second gradient-based algorithm exhibits the additional property that, when stopped, yields a stable approximation with a reduced H2-error norm. We illustrate the theory on a CD-player and on a discretized heat equation.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call