Recurrent neural networks for solving second-order cone programs

  • Authors:
  • Chun-Hsu Ko;Jein-Shan Chen;Ching-Yu Yang

  • Affiliations:
  • Department of Electrical Engineering, I-Shou University, Kaohsiung County 840, Taiwan;Department of Mathematics, National Taiwan Normal University, Taipei 11677, Taiwan;Department of Mathematics, National Taiwan Normal University, Taipei 11677, Taiwan

  • Venue:
  • Neurocomputing
  • Year:
  • 2011

Quantified Score

Hi-index 0.01

Visualization

Abstract

This paper proposes using the neural networks to efficiently solve the second-order cone programs (SOCP). To establish the neural networks, the SOCP is first reformulated as a second-order cone complementarity problem (SOCCP) with the Karush-Kuhn-Tucker conditions of the SOCP. The SOCCP functions, which transform the SOCCP into a set of nonlinear equations, are then utilized to design the neural networks. We propose two kinds of neural networks with the different SOCCP functions. The first neural network uses the Fischer-Burmeister function to achieve an unconstrained minimization with a merit function. We show that the merit function is a Lyapunov function and this neural network is asymptotically stable. The second neural network utilizes the natural residual function with the cone projection function to achieve low computation complexity. It is shown to be Lyapunov stable and converges globally to an optimal solution under some condition. The SOCP simulation results demonstrate the effectiveness of the proposed neural networks.