Optimization with Neural Networks Presented by: Mahmood Khademi Babak Bashiri Instructor: Dr....
-
Upload
florence-lucas -
Category
Documents
-
view
219 -
download
0
Transcript of Optimization with Neural Networks Presented by: Mahmood Khademi Babak Bashiri Instructor: Dr....
Optimization with Neural Networks
Presented by:Mahmood Khademi
Babak BashiriInstructor:
Dr. Bagheri
Sharif University of TechnologyApril 2007
Introduction
An optimization problem consists of two parts: Cost function and Constraints
Constrained The constraints are built in the cost function, so minimizing
the cost function also satisfies the constraints
Unconstraint There is no constraint for the problem!
Combinatorial We separate the constraints and the cost function, minimize
each of them and then add them together
ApplicationApplications in many fields like:
Routing in computer networks VLSI circuit design Planning in operational and logistic systems Power distribution systems Wireless and satellite communication systems
Basic idea
If : decision variables Suppose is our objective function . Constraints can be expressed as nonnegative penalty
functions that only when represent a feasible solution
By combining the penalty functions with F , the original constrained problem may be reformulated as unconstrained problem in which the goal is to minimize the quantity :
nXXX ,...,, 21
)21 ,...,,( nXXXFnXXX ,...,, 21
),...,,( 21 ni XXXC0),...,,( 21 ni XXXC
m
knkn XXXCXXXFF
12121 ),...,,(),...,,(
TSP Is simple to state but very difficult to
solve. The problem is to find the shortest
possible tour through a set of N vertices so that each vertex is visited exactly once.
This problem is known to be NP-complete
Why neural network?
Drawbacks of conventional computing systems: Perform poorly on complex problems Lack the computational power Don’t utilize the inherent parallelism of problems
Advantages of artificial neural networks: Perform well even on complex problems Very fast computational cycles if implemented in hardware Can take the advantage of inherent parallelism of problems
Some Efforts to solve optimization problems
Many ANN algorithms with different architectures have been used to solve different optimization problems…
We’ve selected: Hopfield NN Elastic Net Self Organizing Map NN
Hopfield-Tank model TSP must be mapped, in some way, onto the
neural network structure Each row corresponds to a particular city and
each column to a particular position in the tour
Mapping TSP to Hopfield neural net There is a connection between each pair of
units The signal sent along a connection from i
to t j is equal to the weight Tij if i is activated. It is equal to 0 otherwise.
A negative weight defines inhibitory connection between the two units
It is unlikely that two units with negative weigh will be active or “on” at the same time
Discrete Hopfield Model connection weights are not learned Hopfield network evolves by updating the
activation of each unit in turn In final state, all units are stable according to the
update rule The units are updated at random, one unit at a
time
{Vi}i=1,...,L, L :number of units Vi :activation level of unit iTij: connection weight between units i and jtetai: threshold of unit i.
Discrete Hopfield Model (Cont.)
Energy function
Units changes its activation level if and only if the energy of the network decreases by doing so:
Since the energy can only decrease over time and the number configuration is finite
the network must converge (but not necessarily the minimum energy state)
Continuous Hopfield-Tank Neuron function is
continuous (Sigmoid function)
The evolution of the units over time is now characterized by the following differential equation :
Ui, Ii and Vi are the input, input bias, and activation level of unit I, respectively
Continuous Hopfield-Tank Energy function
Discrete time approximation is applied to the equations of motion
Application of the Hopfield-Tank Model to the TSP
Application of the Hopfield-Tank model to the TSP(1)The TSP is represented as an N*N matrix(2) Energy function
(3)Bias and connection weights are derived
Application of the Hopfield-Tank model to the TSP
Results of Hopfield-Tank Hopfield and Tank were able to solve a
randomly generated 10-city,with parameter value :A=B=500,C=200,N=15.
They reported for 20 trails, network converge 16 times to feasible tours.
Half of those tours were one of two optimal tours
The size of each black square indicates the value of the output of the corresponding neuron
The main weaknesses of the original Hopfield-Tank model
The main weaknesses of the original Hopfield-Tank model(d) Model plagued with the limitation of “hill-
climbing” approaches (e) Model does not guarantee feasibility
The main weaknesses of the original Hopfield-Tank model
The positive points: Can easily implemented in hardware Can be applied to non-Euclidean TSPs
Elastic net (Willshaw-Von der Malsburg)
Elastic net
Energy function for Elastic net
The self organizing map The SOM are instances of “competitive
NN” , used by unsupervised learning system to classify data
Adjusting the weights Related to elastic net Differ of elastic net
Competitive Network Group a set of I-dimensional input pattern
in to K cluster (K<=M)
SOM in the TSP context A set of 2-dimensional coordinates must
be mapped onto a set of 1-dimensional positions in the tour
SOM in the TSP context
Different SOM based on that form Fort increased speed of convergence by reducing neighborhood and reducing modification to weights of neighboring units over time. The work of Angeniol
Questions ?