Скачать презентацию Artificial Neural Networks Rados Jovanovic Summary Скачать презентацию Artificial Neural Networks Rados Jovanovic Summary

80cee77d593b9d91ac7644308ef35b00.ppt

  • Количество слайдов: 29

Artificial Neural Networks Rados Jovanovic Artificial Neural Networks Rados Jovanovic

Summary • • • Biological Model of Neural Networks Mathematical Model of Neural Networks Summary • • • Biological Model of Neural Networks Mathematical Model of Neural Networks Building an Artificial Neural Network Theoretical Properties Applications of Artificial Neural Networks

Biological Model • The brain is responsible for all processing and memory. • The Biological Model • The brain is responsible for all processing and memory. • The brain consists of a complex network of cells called neurons. • Neurons communicate by transimitting electrochemical signals throughout the network. • Each input signal to a neuron can inhibit or excite the neuron. When the neuron is excited enough, it will fire its own electrochemical signal.

Biological Model (Cont. ) A neuron: Biological Model (Cont. ) A neuron:

Mathematical Model • An Artificial Neural Network is a network of interconnected artificial neurons. Mathematical Model • An Artificial Neural Network is a network of interconnected artificial neurons. • Like in a biological neural network, artificial neurons communicate by sending signals to one another. • Each input to an artificial neuron can either inhibit or excite the artificial neuron.

Mathematical Model (Cont. ) An artificial neuron: Mathematical Model (Cont. ) An artificial neuron:

Building an Artificial Neural Network • Topology of the network • Learning type • Building an Artificial Neural Network • Topology of the network • Learning type • Learning algorithm – Backpropagation • Summary of backpropagation

Topology of the Network (Cont. ) There are many topologies, but the main distinction Topology of the Network (Cont. ) There are many topologies, but the main distinction can be maid between: • Feed-Forward Neural Networks • Recurrent Neural Networks

Feed-Forward Neural Networks have no connections that loop. Feed-Forward Neural Networks have no connections that loop.

Recurrent Neural Networks Recurrent neural networks do contain looping connections Recurrent Neural Networks Recurrent neural networks do contain looping connections

Building an Artificial Neural Network • Topology of the network • Learning type • Building an Artificial Neural Network • Topology of the network • Learning type • Learning algorithm – Backpropagation • Summary of backpropagation

Learning Type • Supervised Learning Requires a set of pairs of inputs and outputs Learning Type • Supervised Learning Requires a set of pairs of inputs and outputs to train the artificial neural network on. • Unsupervised Learning Only requires inputs. Through time an ANN learns to organize and cluster data by itself. • Reinforcement Learning An ANN from the given input produces some output, and the ANN is rewarded or punished based on the output it created.

Building an Artificial Neural Network • Topology of the network • Learning type • Building an Artificial Neural Network • Topology of the network • Learning type • Learning algorithm – Backpropagation • Summary of backpropagation

Learning Algorithm Learning is adjustment of the weights of the connections between neurons, according Learning Algorithm Learning is adjustment of the weights of the connections between neurons, according to some modification rule.

Backpropagation • One of the more common algorithms for supervised learning is Backpropagation. • Backpropagation • One of the more common algorithms for supervised learning is Backpropagation. • The term is an abbreviation for “backwards propagation of errors" • Backpropagation is most useful for feedforward networks.

Backpropagation (Cont. ) • An input is fed into the network and the output Backpropagation (Cont. ) • An input is fed into the network and the output is being calculated. • We compare the output of the network with the target output, and we get the error. • We want to minimize the error, so we greedily adjust the weights such that error for this particular input will go towards zero. • We do so using the delta rule.

Delta Rule • The delta rule is a gradient descent learning rule. • For Delta Rule • The delta rule is a gradient descent learning rule. • For a given neuron j and a weight i the delta rule for its weight wji is: • • tj is the target output yj is the actual output xi is the ith input a is a small constant called the learning rate

Delta Rule (Cont. ) Delta Rule (Cont. )

Delta Rule (Cont. ) Delta Rule (Cont. )

Backpropagation (Cont. ) • The whole process is repeated for each of the training Backpropagation (Cont. ) • The whole process is repeated for each of the training cases, then back to the first case again. • The cycle is repeated until the overall error value drops below some pre-determined threshold. • Backpropagation usually allows quick convergence on satisfactory local minima for error.

Summary of Backpropagation -While error over all training samples > threshold -For each training Summary of Backpropagation -While error over all training samples > threshold -For each training sample: -Present a training sample to the neural network. -Compute the output of the network. -Calculate the error for each neuron in the last layer by comparing it to the target output. -For each layer l starting from the last output layer until the first layer -Adjust the weights for each neuron in layer l to decrease the local error. -Propagate the error to each neuron one layer back -end for

Theoretical Properties • • Computational power Capacity Generalisation Convergence Theoretical Properties • • Computational power Capacity Generalisation Convergence

Theoretical Properties • Computational power: The Cybenko theorem proved single hidden layer, feed forward Theoretical Properties • Computational power: The Cybenko theorem proved single hidden layer, feed forward neural network is capable of approximating any continuous, multivariate function to any desired degree of accuracy.

Theoretical Properties (Cont. ) • Capacity: It roughly corresponds to the neural network’s ability Theoretical Properties (Cont. ) • Capacity: It roughly corresponds to the neural network’s ability to model any given function. It is related to the amount of information that can be stored in the network.

Theoretical Properties (Cont. ) • Generalisation: In applications where the goal is to create Theoretical Properties (Cont. ) • Generalisation: In applications where the goal is to create a system that generalises well in unseen examples, the problem of overtraining has emerged. • To lessen the overtraining cross validation is used.

Theoretical Properties (Cont. ) • Convergence: Not much can be said about convergence since Theoretical Properties (Cont. ) • Convergence: Not much can be said about convergence since it depends on many factors, such as the existence of local minima, choice of optimization method, etc.

Applications Application areas include: system identification and control (vehicle control, process control), quantum chemistry, Applications Application areas include: system identification and control (vehicle control, process control), quantum chemistry, gameplaying and decision making (backgammon, chess, racing), pattern recognition (radar systems, face identification, object recognition. . . ), sequence recognition (gesture, speech, handwritten text recognition), medical diagnosis, financial applications (automated trading systems), data mining, visualization, e-mail spam filtering. . .

Fin Questions? Fin Questions?