Recurrent neural networks for solving second-order cone programs

Chun Hsu Ko, Jein Shan Chen*, Ching Yu Yang

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

29 Citations (Scopus)


This paper proposes using the neural networks to efficiently solve the second-order cone programs (SOCP). To establish the neural networks, the SOCP is first reformulated as a second-order cone complementarity problem (SOCCP) with the Karush-Kuhn-Tucker conditions of the SOCP. The SOCCP functions, which transform the SOCCP into a set of nonlinear equations, are then utilized to design the neural networks. We propose two kinds of neural networks with the different SOCCP functions. The first neural network uses the Fischer-Burmeister function to achieve an unconstrained minimization with a merit function. We show that the merit function is a Lyapunov function and this neural network is asymptotically stable. The second neural network utilizes the natural residual function with the cone projection function to achieve low computation complexity. It is shown to be Lyapunov stable and converges globally to an optimal solution under some condition. The SOCP simulation results demonstrate the effectiveness of the proposed neural networks.

Original languageEnglish
Pages (from-to)3646-3653
Number of pages8
Issue number17
Publication statusPublished - 2011 Oct


  • Cone projection function
  • Fischer-Burmeister function
  • Lyapunov stable
  • Merit function
  • Neural network
  • SOCP

ASJC Scopus subject areas

  • Computer Science Applications
  • Cognitive Neuroscience
  • Artificial Intelligence


Dive into the research topics of 'Recurrent neural networks for solving second-order cone programs'. Together they form a unique fingerprint.

Cite this