next up previous index
Next: RL using Lagrange-Hopfield Method Up: Constrained Minimization Previous: Lagrange Multipliers

8.3.3

Hopfield Method

 

The graded Hopfield neural network (HNN) [Hopfield 1984] provides another way of applying penalty functions. Let us use the constrained minimization in RL as an example to illustrate this point. There, the feasibility (8.19) must be satisfied by the labeling assignment at any time and the unambiguity (8.22) should be satisfied by the final labeling assignment. The unambiguity constraint can be imposed by the following term

where is the inverse of a function to be illustrated below. A local state is related to an internal variable by

where is usually a sigmoid function

controlled by the parameter T. In very high gain when , is forced to be 0 or 1 depending on whether is positive or negative [Hopfield 1984]. The term reaches the minimum of zero only when all 's are either 0 or 1. This means that minimizing this term with in effect leads to unambiguous labeling.

On the other hand, The feasibility can be imposed by

This term has its minimum value of zero when the feasibility is satisfied. Now the constrained optimization minimizes the following functional

where a and b are weights. A minimal solution which is feasible and unambiguous is given by

To derive a dynamic system for the minimization, introduce a time variable into f such that . The energy change due to state change is

where

is a gradient component. The following dynamics minimizes the three term energy :

where capacitance C>0 controls the convergence of the system. With the above dynamics, the energy change

is non-positive since is a monotonically increasing function and C is positive. This updating rule will lead to a minimal solution which is both feasible and unambiguous when , and .

To solve combinatorial optimization, Wacholder et al. (1989) develop an algorithm which combines the HNN of [Hopfield 1984] and the Lagrange multiplier dynamics of [Platt and Barr 1988]. Good convergence to feasible solutions is reported. In Section 8.3.4, we will describe a method which combines the HNN and the augmented Lagrangian function to perform relaxation labeling.

The HNN approach has been used to convert binary-valued line process variables to real ones for low level process [Koch et al. 1986,Yuille 1987] ( cf. Section sec:DA.LP). It has also been used as an alternative to relaxation labeling for matching; this is seen as symbolic image labeling [Jamison and Schalkoff 1988], subgraph isomorphism for object recognition [Nasrabadi et al. 1990 ], and matching to multiple view models [Lin et al, 1991].

Comparing RL and HNN for combinatorial minimization like object matching, our experience is that the former performs better. Due to the influence of the extra penalty function added in the HNN approach, the modified energy may no longer solve the original matching problem [Horn 1988]. Moreover, it easily leads to unfavorable local solutions. The RL algorithm is much more careful in this regard. It uses gradient projection to choose the best direction and magnitude for the state to evolve. In other words, the problem of local optima is more significant in the HNN approach. Computationally, the RL needs a smaller number of iterations before convergence.



next up previous index
Next: RL using Lagrange-Hopfield Up: Constrained Minimization Previous: Lagrange Multipliers