Abstract

Recently, several studies have proposed methods to utilize some classes of optimization problems in designing deep neural networks to encode constraints that conventional layers cannot capture. However, these methods are still in their infancy and require special treatments, such as the analysis of the Karush–Kuhn–Tucker (KKT) condition, to derive the backpropagation formula. In this paper, we propose a new formulation called the fixed-point iteration (FPI) layer, which facilitates the use of more complicated operations in deep networks. The backward FPI layer, which is motivated by the recurrent backpropagation (RBP) algorithm, is also proposed. However, in contrast to RBP, the backward FPI layer yields the gradient using a small network module without explicitly calculating the Jacobian. In actual applications, both forward and backward FPI layers can be treated as nodes in the computational graphs. All the components of our method are implemented at a high level of abstraction, which allows efficient higher-order differentiations on the nodes. In addition, we present two practical methods, the neural net FPI (FPI_NN) layer and the gradient descent FPI (FPI_GD) layer, whereby the FPI update operations are a small neural network module and a single gradient descent step based on a learnable cost function, respectively. FPI_NN is intuitive and simple, while FPI_GD can be used to efficient train energy function networks that have been studied recently. While RBP and related studies have not been applied to practical examples, our experiments show that the FPI layer can be successfully applied to real-world problems such as image denoising, optical flow, and multi-label classification.

Highlights

  • Several papers have proposed the composition of deep neural network with more complicated algorithms rather than the simple operations that have been used so far

  • We employ the fixed-point iteration (FPI), which is the basis for many numerical algorithms, including most gradient-based optimizations, as a layer in the neural network

  • We propose a backpropagation method called the backward FPI layer based on recurrent backpropagation (RBP) [3], [22] to efficiently compute the derivative of the FPI layer

Read more

Summary

INTRODUCTION

Several papers have proposed the composition of deep neural network with more complicated algorithms rather than the simple operations that have been used so far. To facilitate the use of more complicated operations in deep networks, in this paper, we introduce a new layer formulation that can be practically implemented and trained based on RBP with some additional considerations. To this end, we employ the fixed-point iteration (FPI), which is the basis for many numerical algorithms, including most gradient-based optimizations, as a layer in the neural network. We provide a modularized implementation of the partial differentiation operation, which is essential in the backward FPI layer but absent from regular autograd libraries, based on an independent computational graph This makes the proposed method very convenient for various practical applications.

AND RELATED WORK
STRUCTURE OF THE FIXED-POINT ITERATION LAYER
DIFFERENTIATION OF THE FPI LAYER
BACKWARD FIXED-POINT ITERATION LAYER
CONVERGENCE OF THE FPI LAYER
EXPERIMENTS
Findings
CONCLUSION AND FUTURE WORK
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.