Approximate gradient/penalty methods with general discretization schemes for optimal control problems

  • Authors:
  • Ion Chryssoverghi

  • Affiliations:
  • Department of Mathematics, National Technical University of Athens, Athens, Greece

  • Venue:
  • LSSC'05 Proceedings of the 5th international conference on Large-Scale Scientific Computing
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

We consider an optimal control problem described by ordinary differential equations, with control and state constraints. The state equation is first discretized by a general explicit Runge-Kutta scheme and the controls are approximated by piecewise polynomial functions. We then propose approximate gradient and gradient projection methods, and their penalized versions, that construct sequences of discrete controls and progressively refine the discretization during the iterations. Instead of using the exact discrete cost derivative, which usually requires tedious calculations of composite functions, we use here an approximate derivative of the cost defined by discretizing the continuous adjoint equation by the same, but nonmatching, Runge-Kutta scheme backward and the integral involved by a Newton-Cotes integration rule. We show that strong accumulation points in L2 of sequences constructed by these methods satisfy the weak necessary conditions for optimality for the continuous problem. Finally, numerical examples are given.