Monotonicity recovering and accuracy preserving optimization methods for postprocessing finite element solutions

  • Authors:
  • Oleg Burdakov;Ivan Kapyrin;Yuri Vassilevski

  • Affiliations:
  • Department of Mathematics, Linköping University, Linköping, Sweden;Institute of Numerical Mathematics, Russian Academy of Sciences, Moscow, Russia;Institute of Numerical Mathematics, Russian Academy of Sciences, Moscow, Russia

  • Venue:
  • Journal of Computational Physics
  • Year:
  • 2012

Quantified Score

Hi-index 31.45

Visualization

Abstract

We suggest here a least-change correction to available finite element (FE) solution. This postprocessing procedure is aimed at recovering the monotonicity and some other important properties that may not be exhibited by the FE solution. Although our approach is presented for FEs, it admits natural extension to other numerical schemes, such as finite differences and finite volumes. For the postprocessing, a priori information about the monotonicity is assumed to be available, either for the whole domain or for a subdomain where the lost monotonicity is to be recovered. The obvious requirement is that such information is to be obtained without involving the exact solution, e.g. from expected symmetries of this solution. The postprocessing is based on solving a monotonic regression problem with some extra constraints. One of them is a linear equality-type constraint that models the conservativity requirement. The other ones are box-type constraints, and they originate from the discrete maximum principle. The resulting postprocessing problem is a large scale quadratic optimization problem. It is proved that the postprocessed FE solution preserves the accuracy of the discrete FE approximation. We introduce an algorithm for solving the postprocessing problem. It can be viewed as a dual ascent method based on the Lagrangian relaxation of the equality constraint. We justify theoretically its correctness. Its efficiency is demonstrated by the presented results of numerical experiments.