尊敬的 微信汇率:1円 ≈ 0.046078 元 支付宝汇率:1円 ≈ 0.046168元 [退出登录]
SlideShare a Scribd company logo
Neural Networks
Dr. Randa Elanwar
Lecture 9
Lecture Content
• Mapping networks:
– Back-propagation neural network
– Self-organizing map
– Counter propagation network
• Spatiotemporal Network
• Stochastic Networks
– Boltzmann machine
• Neurocognition network
2Neural Networks Dr. Randa Elanwar
Mapping networks
• When the problem is non linear and no straight line
could ever separate samples in the feature space we
need multilayer perceptrons (having hidden layer’s’) to
achieve nonlinearity.
• The idea is that we map/transform/translate our data
to another feature space that is linearly separable. Thus
we call them mapping networks.
• We will discuss three types of mapping networks: the
back-propagation neural network, self-organizing map,
counter propagation network.
3Neural Networks Dr. Randa Elanwar
Mapping networks
• Networks without hidden units are very limited in the input-output
mappings they can model.
– More layers of linear units do not help. Its still linear.
– Fixed output non-linearities are not enough
• We need multiple layers of adaptive non-linear hidden units.
• But how can we train such nets?
– We need an efficient way of adapting all the weights, not just the
last layer. i.e., Learning the weights going into hidden units . This is
hard.
– Why?
– Because: Nobody is telling us directly what hidden units should
do.
– Solution: This can be achieved using ‘Backpropagation’ learning
4Neural Networks Dr. Randa Elanwar
Learning with hidden layers
• Mathematically, the learning process is an optimization problem. We
initiate the NN system with some parameters (weights) and use
known examples to find out the optimal values of such weights.
• Generally, the solution of an optimization problem is to find the
parameter value that leads to minimum value of an optimization
function.
5Neural Networks Dr. Randa Elanwar
G(t)
t
In our case, the optimization function that we
need to minimize to get the final weights is the
error function
E = ydes-yact
E = ydes-f(W.X)
To get the minimum value mathematically we
differentiate the error function with respect to
the parameter we need to get we call it W
E


Learning with hidden layers
• We define the “gradient”: w = . . X
• If  is +ve this means that the current values of W makes the
differentiation result +ve which is wrong. We want
differentiation result to be = 0 (minimum point) we must
move in the opposite direction of the gradient (subtract). The
opposite is also true.
• If  is =0 this means that the current values of W makes the
differentiation result = 0 which is right. These weights are the
optimal values (solution) and w should stop the algorithm. The
network now is trained and ready for use.
6Neural Networks Dr. Randa Elanwar
The back propagation algorithm
• The backpropagation learning algorithm can be divided into two
phases: propagation and weight update.
7Neural Networks Dr. Randa Elanwar
Phase 1: Propagation
1.Forward propagation of a training
pattern's input through the neural
network in order to generate the
propagation's output activations
(yact).
2.Backward propagation of the
propagation's output activations
through the neural network using
the training pattern's target (ydes) in
order to generate the deltas () of
all output and hidden neurons.
Phase 2: Weight update
For each weight follow the following steps:
1.Multiply its output delta () and input
activation (x) and the learning rate () to
get the gradient of the weight (w).
2.Bring the weight in the opposite direction
of the gradient by subtracting it from the
weight.
- The sign of the gradient of a weight
indicates where the error is increasing, this
is why the weight must be updated in the
opposite direction.
- Repeat phase 1 and 2 until the
performance of the network is satisfactory.
Backpropagation Networks
• They are the nonlinear (mapping) neural networks using the
backpropagation supervised learning technique.
• Modes of learning of nonlinear nets:
• There are three modes of learning to choose from: on-line
(pattern), batch and stochastic.
• In on-line and stochastic learning, each propagation is followed
immediately by a weight update.
• In batch learning, many propagations occur before updating the
weights.
• Batch learning requires more memory capacity, but on-line and
stochastic learning require more updates.
8Neural Networks Dr. Randa Elanwar
Backpropagation Networks
• On-line learning is used for dynamic environments that
provide a continuous stream of new patterns.
• Stochastic learning and batch learning both make use
of a training set of static patterns. Stochastic goes
through the data set in a random order in order to
reduce its chances of getting stuck in local minima.
• Stochastic learning is also much faster than batch
learning since weights are updated immediately after
each propagation. Yet batch learning will yield a much
more stable descent to a local minima since each
update is performed based on all patterns.
9Neural Networks Dr. Randa Elanwar
Backpropagation Networks
• Applications of supervised learning (Backpropagation NN)
include
• Pattern recognition
• Credit approval
• Target marketing
• Medical diagnosis
• Defective parts identification in manufacturing
• Crime zoning
• Treatment effectiveness analysis
• Etc
10Neural Networks Dr. Randa Elanwar
Self-organizing map
• We can also train networks where there is no teacher. This is called
unsupervised learning. The network learns a prototype based on the
distribution of patterns in the training data. Such networks allow us
to:
– Discover underlying structure of the data
– Encode or compress the data
– Transform the data
• Self-organizing maps (SOMs) are a data visualization technique
invented by Professor Teuvo Kohonen
– Also called Kohonen Networks, Competitive Learning, Winner-Take-All
Learning
– Generally reduces the dimensions of data through the use of self-
organizing neural networks
– Useful for data visualization; humans cannot visualize high dimensional
data so this is often a useful technique to make sense of large data sets
11Neural Networks Dr. Randa Elanwar
Self-organizing map
• SOM structure:
1. Weights in neuron must represent a class of
pattern. We have a neuron for each class.
2. Inputs pattern presented to all neurons and each
produces an output. Output: measure of the match
between input pattern and pattern stored by
neuron.
3. A competitive learning strategy selects neuron with
largest response.
4. A method of reinforcing the largest response.
12Neural Networks Dr. Randa Elanwar
Self-organizing map
• Unsupervised classification learning is based on clustering of
input data. No a priori knowledge is about an input’s
membership in a particular class.
• Instead, gradually detected characteristics and a history of
training will be used to assist the network in defining classes
and possible boundaries between them.
• Clustering is understood to be the grouping of similar objects
and separating of dissimilar ones.
• We discuss Kohonen’s network which classifies input vectors
into one of the specified number of m categories, according to
the clusters detected in the training set
13Neural Networks Dr. Randa Elanwar
Kohonen’s Network
14Neural Networks Dr. Randa Elanwar
Kohonen network
X
•The Kohonen network is a self-organising
network with the following
characteristics:
1. Neurons are arranged on a 2D grid
2. Inputs are sent to all neurons
3. There are no connections between
neurons
4. For a neuron output (j) is a weighted
sum of multiplication of x and w
vectors, where x is the input, w is the
weights
5. There is no threshold or bias
6. Input values and weights are
normalized
Self-organizing map
Learning in Kohonen networks:
• Initially the weights in each neuron are random
• Input values are sent to all the neurons
• The outputs of each neuron are compared
• The “winner” is the neuron with the largest output value
• Having found the winner, the weights of the winning neuron are
adjusted
• Weights of neurons in a surrounding neighbourhood are also
adjusted
• As training progresses the neighbourhood gets smaller
• Weights are adjusted according to the following formula:
15Neural Networks Dr. Randa Elanwar
Self-organizing map
• The learning coefficient (alpha) starts with a value of 1 and
gradually reduces to 0
• This has the effect of making big changes to the weights initially,
but no changes at the end
• The weights are adjusted so that they more closely resemble
the input patterns
Applications of unsupervised learning (Kohonen’s NN) include
• Clustering
• Vector quantization
• Data compression
• Feature extraction
16Neural Networks Dr. Randa Elanwar
Counter propagation network
• The counterpropagation network (CPN) is a fast-learning
combination of unsupervised and supervised learning.
• Although this network uses linear neurons, it can learn nonlinear
functions by means of a hidden layer of competitive units.
• Moreover, the network is able to learn a function and its inverse
at the same time.
• However, to simplify things, we will only consider the
feedforward mechanism of the CPN.
17Neural Networks Dr. Randa Elanwar
Counter propagation network
• Training:
1.Randomly select a vector pair (x, y) from the training set.
2.Measure the similarity between the input vector and the
activation of the hidden-layer units.
3.In the hidden (competitive) layer, determine the unit with the
largest activation (the winner). I.e., the neuron whose weight
vector is most similar to the current input vector is the “winner.”
4.Adjust the connection weights inbetween
5.Repeat until each input pattern is consistently associated with
the same competitive unit.
18Neural Networks Dr. Randa Elanwar
Counter propagation network
• After the first phase of the training, each hidden-layer neuron is
associated with a subset of input vectors (class of patterns).
• In the second phase of the training, we adjust the weights in the
network’s output layer in such a way that, for any winning hidden-
layer unit, the network’s output is as close as possible to the desired
output for the winning unit’s associated input vectors.
• The idea is that when we later use the network to compute
functions, the output of the winning hidden-layer unit is 1, and the
output of all other hidden-layer units is 0.
19Neural Networks Dr. Randa Elanwar
Spatiotemporal Networks
•A spatio-temporal neural net differs from other neural networks in two ways:
1. Neurons has recurrent links that have different propagation delays
2. The state of the network depends not only on which nodes are firing, but
also on the relative firing times of nodes. i.e., the significance of a node
varies with time and depends on the firing state of other nodes.
•The use of recurrence and multiple links with variable propagation delays
provides a rich mechanism for feature extraction and pattern recognition:
1. Recurrent links enable nodes to integrate and differentiate inputs. I.e.,
detect features
2. multiple links with variable propagation delays between nodes serve as a
short-term memory.
20Neural Networks Dr. Randa Elanwar
Spatiotemporal Networks
• Applications:
• Problems such as speech recognition and time series prediction where the
input signal has an explicit temporal aspect.
• Tasks like image recognition do not have an explicit temporal aspect, but
can also be done by converting static patterns into time-varying (spatio-
temporal) signals via scanning the image. This would lead to a number of
significant advantages:
– The recognition system becomes ‘shift invarient’
– The spatio-temporal approach explains the image geometry since the local
spatial relationships in the image are expressed as local temporal variations in
the scanned input.
– Reduction of complexity (from 2D to 1D)
– The scanning approach allows a visual pattern recognition system to deal with
inputs of arbitrary extent (not only static fixed 2D pattern)
21Neural Networks Dr. Randa Elanwar
Stochastic neural networks
• Stochastic neural networks are a type of artificial neural
networks, which is a tool of artificial intelligence. They are
built by introducing random variations into the network,
either by giving the network's neurons stochastic transfer
functions, or by giving them stochastic weights. This makes
them useful tools for optimization problems, since the
random fluctuations help it escape from local minima.
• Stochastic neural networks that are built by using stochastic
transfer functions are often called Boltzmann machines.
• Stochastic neural networks have found applications in risk
management, oncology, bioinformatics, and other similar
fields
22Neural Networks Dr. Randa Elanwar
Stochastic Networks: Boltzmann machine
• The neurons are stochastic: at any time there is a probability
attached to whether the neurons fires.
• Used for solving constrained optimization problems.
• Typical Boltzmann Machine:
– Weights are fixed to represent the constrains of the problem and the
function to be optimized.
– The net seeks the solution by changing the activations of the units (0 or
1) based on a probability distribution and the effect that the change
would have on the energy function or consensus function for the net.
• May use either supervised or unsupervised learning.
• Learning in Boltzmann Machine is accomplished by using a
Simulated Annealing technique which has stochastic nature. This is
used to reduce the probability of the net becoming trapped in a
local minimum which is not a global minimum.
23Neural Networks Dr. Randa Elanwar
Stochastic Networks: Boltzmann machine
• Learning characteristics:
– Each neuron fires with bipolar values.
– All connections are symmetric.
– In activation passing, the next neuron whose state we
wish to update is selected randomly.
– There are no self-feedback (connections from a neuron
to itself)
24Neural Networks Dr. Randa Elanwar
Stochastic Networks: Boltzmann machine
• There are three phases in operation of the network:
– The clamped phase in which the input and output of visible
neurons are held fixed, while the hidden neurons are allowed
to vary.
– The free running phase in which only the inputs are held fixed
and other neurons are allowed to vary.
– The learning phase.
• These phases iterate till learning has created a
Boltzmann Machine which can be said to have learned
the input patterns and will converge to the learned
patterns when noisy or incomplete pattern is
presented.
25Neural Networks Dr. Randa Elanwar
Stochastic Networks: Boltzmann machine
• For unsupervised learning Generally the initial weights of the net
are randomly set to values in a small range e.g. -0.5 to +0.5.
• Then an input pattern is presented to the net and clamped to the
visible neurons.
• choose a hidden neurons at random and flip its state from sj to –sj
according to certain probability distribution
• The activation passing can continue till the net hidden neurons
reach equilibrium.
• During free running phase, after presentation of the input patterns
all neurons can update their states.
• The learning phase depends whether weight are changed depend
on the difference between the "real" distribution (neuron state) in
clamped phase and the one which will be produced (eventually) by
the machine in free mode.
26Neural Networks Dr. Randa Elanwar
Stochastic Networks: Boltzmann machine
• For supervised learning the set of visible neurons is split into
input and output neurons, and the machine will be used to
associate an input pattern with an output pattern.
• During the clamped phase, the input and output patterns are
clamped to the appropriate units.
• The hidden neurons’ activations can settle at the various
values.
• During free running phase, only the input neurons are
clamped – both the output neurons and the hidden neurons
can pass activation round till the activations in the network
settles.
• Learning rule here is the same as before but must be
modulated (multiplied) by the probability of the input’s
patterns
27Neural Networks Dr. Randa Elanwar
Neurocognition network
• Neurocognitive networks are large-scale systems of
distributed and interconnected neuronal populations in
the central nervous system organized to perform
cognitive functions.
• many computer scientists try to simulate human
cognition with computers. This line of research can be
roughly split into two types: research seeking to create
machines as adept as humans (or more so), and
research attempting to figure out the computational
basis of human cognition — that is, how the brain
actually carries out its computations. This latter branch
of research can be called computational modeling
(while the former is often called artificial intelligence or
AI).
28Neural Networks Dr. Randa Elanwar

More Related Content

What's hot

Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
Knoldus Inc.
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
sweetysweety8
 
Artificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rulesArtificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rules
Mohammed Bennamoun
 
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNSArtificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
Mohammed Bennamoun
 
Introduction to Neural networks (under graduate course) Lecture 1 of 9
Introduction to Neural networks (under graduate course) Lecture 1 of 9Introduction to Neural networks (under graduate course) Lecture 1 of 9
Introduction to Neural networks (under graduate course) Lecture 1 of 9
Randa Elanwar
 
Lecture artificial neural networks and pattern recognition
Lecture   artificial neural networks and pattern recognitionLecture   artificial neural networks and pattern recognition
Lecture artificial neural networks and pattern recognition
Hưng Đặng
 
Nural network ER. Abhishek k. upadhyay
Nural network ER. Abhishek  k. upadhyayNural network ER. Abhishek  k. upadhyay
Nural network ER. Abhishek k. upadhyay
abhishek upadhyay
 
Ann
Ann Ann
Ann
vini89
 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
NikitaRuhela
 
03 Single layer Perception Classifier
03 Single layer Perception Classifier03 Single layer Perception Classifier
03 Single layer Perception Classifier
Tamer Ahmed Farrag, PhD
 
Introduction to Neural networks (under graduate course) Lecture 2 of 9
Introduction to Neural networks (under graduate course) Lecture 2 of 9Introduction to Neural networks (under graduate course) Lecture 2 of 9
Introduction to Neural networks (under graduate course) Lecture 2 of 9
Randa Elanwar
 
Artificial neural network model & hidden layers in multilayer artificial neur...
Artificial neural network model & hidden layers in multilayer artificial neur...Artificial neural network model & hidden layers in multilayer artificial neur...
Artificial neural network model & hidden layers in multilayer artificial neur...
Muhammad Ishaq
 
Artificial neural networks (2)
Artificial neural networks (2)Artificial neural networks (2)
Artificial neural networks (2)
sai anjaneya
 
Neural network
Neural networkNeural network
Neural network
KRISH na TimeTraveller
 
Artificial neural networks seminar presentation using MSWord.
Artificial neural networks seminar presentation using MSWord.Artificial neural networks seminar presentation using MSWord.
Artificial neural networks seminar presentation using MSWord.
Mohd Faiz
 
Fundamental, An Introduction to Neural Networks
Fundamental, An Introduction to Neural NetworksFundamental, An Introduction to Neural Networks
Fundamental, An Introduction to Neural Networks
Nelson Piedra
 
Artificial Neural Networks for NIU
Artificial Neural Networks for NIUArtificial Neural Networks for NIU
Artificial Neural Networks for NIU
Prof. Neeta Awasthy
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
DEEPASHRI HK
 
02 Fundamental Concepts of ANN
02 Fundamental Concepts of ANN02 Fundamental Concepts of ANN
02 Fundamental Concepts of ANN
Tamer Ahmed Farrag, PhD
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
Priyank Panchmiya
 

What's hot (20)

Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
 
Artificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rulesArtificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rules
 
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNSArtificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
 
Introduction to Neural networks (under graduate course) Lecture 1 of 9
Introduction to Neural networks (under graduate course) Lecture 1 of 9Introduction to Neural networks (under graduate course) Lecture 1 of 9
Introduction to Neural networks (under graduate course) Lecture 1 of 9
 
Lecture artificial neural networks and pattern recognition
Lecture   artificial neural networks and pattern recognitionLecture   artificial neural networks and pattern recognition
Lecture artificial neural networks and pattern recognition
 
Nural network ER. Abhishek k. upadhyay
Nural network ER. Abhishek  k. upadhyayNural network ER. Abhishek  k. upadhyay
Nural network ER. Abhishek k. upadhyay
 
Ann
Ann Ann
Ann
 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
 
03 Single layer Perception Classifier
03 Single layer Perception Classifier03 Single layer Perception Classifier
03 Single layer Perception Classifier
 
Introduction to Neural networks (under graduate course) Lecture 2 of 9
Introduction to Neural networks (under graduate course) Lecture 2 of 9Introduction to Neural networks (under graduate course) Lecture 2 of 9
Introduction to Neural networks (under graduate course) Lecture 2 of 9
 
Artificial neural network model & hidden layers in multilayer artificial neur...
Artificial neural network model & hidden layers in multilayer artificial neur...Artificial neural network model & hidden layers in multilayer artificial neur...
Artificial neural network model & hidden layers in multilayer artificial neur...
 
Artificial neural networks (2)
Artificial neural networks (2)Artificial neural networks (2)
Artificial neural networks (2)
 
Neural network
Neural networkNeural network
Neural network
 
Artificial neural networks seminar presentation using MSWord.
Artificial neural networks seminar presentation using MSWord.Artificial neural networks seminar presentation using MSWord.
Artificial neural networks seminar presentation using MSWord.
 
Fundamental, An Introduction to Neural Networks
Fundamental, An Introduction to Neural NetworksFundamental, An Introduction to Neural Networks
Fundamental, An Introduction to Neural Networks
 
Artificial Neural Networks for NIU
Artificial Neural Networks for NIUArtificial Neural Networks for NIU
Artificial Neural Networks for NIU
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
 
02 Fundamental Concepts of ANN
02 Fundamental Concepts of ANN02 Fundamental Concepts of ANN
02 Fundamental Concepts of ANN
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
 

Viewers also liked

Improving Performance of Back propagation Learning Algorithm
Improving Performance of Back propagation Learning AlgorithmImproving Performance of Back propagation Learning Algorithm
Improving Performance of Back propagation Learning Algorithm
ijsrd.com
 
The Back Propagation Learning Algorithm
The Back Propagation Learning AlgorithmThe Back Propagation Learning Algorithm
The Back Propagation Learning Algorithm
ESCOM
 
Neural network & its applications
Neural network & its applications Neural network & its applications
Neural network & its applications
Ahmed_hashmi
 
AI Lesson 39
AI Lesson 39AI Lesson 39
AI Lesson 39
Assistant Professor
 
Setting Artificial Neural Networks parameters
Setting Artificial Neural Networks parametersSetting Artificial Neural Networks parameters
Setting Artificial Neural Networks parameters
Madhumita Tamhane
 
2.3.1 properties of functions
2.3.1 properties of functions2.3.1 properties of functions
2.3.1 properties of functions
Northside ISD
 
Introduction to Neural networks (under graduate course) Lecture 6 of 9
Introduction to Neural networks (under graduate course) Lecture 6 of 9Introduction to Neural networks (under graduate course) Lecture 6 of 9
Introduction to Neural networks (under graduate course) Lecture 6 of 9
Randa Elanwar
 
Your amazing brain assembly
Your amazing brain assemblyYour amazing brain assembly
Your amazing brain assembly
HighbankPrimary
 
Neuron Mc Culloch Pitts dan Hebb
Neuron Mc Culloch Pitts dan HebbNeuron Mc Culloch Pitts dan Hebb
Neuron Mc Culloch Pitts dan Hebb
Sherly Uda
 
Artificial Neural Network Topology
Artificial Neural Network TopologyArtificial Neural Network Topology
Artificial Neural Network Topology
Harshana Madusanka Jayamaha
 
Neural networks introduction
Neural networks introductionNeural networks introduction
Neural networks introduction
آيةالله عبدالحكيم
 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
R A Akerkar
 
Soft computing01
Soft computing01Soft computing01
Soft computing01
university of sargodha
 
An Introduction to Neural Networks and Machine Learning
An Introduction to Neural Networks and Machine LearningAn Introduction to Neural Networks and Machine Learning
An Introduction to Neural Networks and Machine Learning
Chris Nicholls
 
hopfield neural network
hopfield neural networkhopfield neural network
hopfield neural network
Abhishikha Sinha
 
Back propagation network
Back propagation networkBack propagation network
Back propagation network
HIRA Zaidi
 
Introduction to Neural Networks in Tensorflow
Introduction to Neural Networks in TensorflowIntroduction to Neural Networks in Tensorflow
Introduction to Neural Networks in Tensorflow
Nicholas McClure
 
Hopfield Networks
Hopfield NetworksHopfield Networks
Hopfield Networks
Kanchana Rani G
 
backpropagation in neural networks
backpropagation in neural networksbackpropagation in neural networks
backpropagation in neural networks
Akash Goel
 
HOPFIELD NETWORK
HOPFIELD NETWORKHOPFIELD NETWORK
HOPFIELD NETWORK
ankita pandey
 

Viewers also liked (20)

Improving Performance of Back propagation Learning Algorithm
Improving Performance of Back propagation Learning AlgorithmImproving Performance of Back propagation Learning Algorithm
Improving Performance of Back propagation Learning Algorithm
 
The Back Propagation Learning Algorithm
The Back Propagation Learning AlgorithmThe Back Propagation Learning Algorithm
The Back Propagation Learning Algorithm
 
Neural network & its applications
Neural network & its applications Neural network & its applications
Neural network & its applications
 
AI Lesson 39
AI Lesson 39AI Lesson 39
AI Lesson 39
 
Setting Artificial Neural Networks parameters
Setting Artificial Neural Networks parametersSetting Artificial Neural Networks parameters
Setting Artificial Neural Networks parameters
 
2.3.1 properties of functions
2.3.1 properties of functions2.3.1 properties of functions
2.3.1 properties of functions
 
Introduction to Neural networks (under graduate course) Lecture 6 of 9
Introduction to Neural networks (under graduate course) Lecture 6 of 9Introduction to Neural networks (under graduate course) Lecture 6 of 9
Introduction to Neural networks (under graduate course) Lecture 6 of 9
 
Your amazing brain assembly
Your amazing brain assemblyYour amazing brain assembly
Your amazing brain assembly
 
Neuron Mc Culloch Pitts dan Hebb
Neuron Mc Culloch Pitts dan HebbNeuron Mc Culloch Pitts dan Hebb
Neuron Mc Culloch Pitts dan Hebb
 
Artificial Neural Network Topology
Artificial Neural Network TopologyArtificial Neural Network Topology
Artificial Neural Network Topology
 
Neural networks introduction
Neural networks introductionNeural networks introduction
Neural networks introduction
 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
 
Soft computing01
Soft computing01Soft computing01
Soft computing01
 
An Introduction to Neural Networks and Machine Learning
An Introduction to Neural Networks and Machine LearningAn Introduction to Neural Networks and Machine Learning
An Introduction to Neural Networks and Machine Learning
 
hopfield neural network
hopfield neural networkhopfield neural network
hopfield neural network
 
Back propagation network
Back propagation networkBack propagation network
Back propagation network
 
Introduction to Neural Networks in Tensorflow
Introduction to Neural Networks in TensorflowIntroduction to Neural Networks in Tensorflow
Introduction to Neural Networks in Tensorflow
 
Hopfield Networks
Hopfield NetworksHopfield Networks
Hopfield Networks
 
backpropagation in neural networks
backpropagation in neural networksbackpropagation in neural networks
backpropagation in neural networks
 
HOPFIELD NETWORK
HOPFIELD NETWORKHOPFIELD NETWORK
HOPFIELD NETWORK
 

Similar to Introduction to Neural networks (under graduate course) Lecture 9 of 9

Unit 2 ml.pptx
Unit 2 ml.pptxUnit 2 ml.pptx
Unit 2 ml.pptx
PradeeshSAI
 
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTS
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTSNEURAL NETWORK IN MACHINE LEARNING FOR STUDENTS
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTS
hemasubbu08
 
Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...
Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...
Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...
Simplilearn
 
Artificial Neural Network (ANN
Artificial Neural Network (ANNArtificial Neural Network (ANN
Artificial Neural Network (ANN
Andrew Molina
 
Neural network
Neural networkNeural network
Neural network
Saddam Hussain
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
IshaneeSharma
 
Introduction to Perceptron and Neural Network.pptx
Introduction to Perceptron and Neural Network.pptxIntroduction to Perceptron and Neural Network.pptx
Introduction to Perceptron and Neural Network.pptx
Poonam60376
 
Terminology Machine Learning
Terminology Machine LearningTerminology Machine Learning
Terminology Machine Learning
DataminingTools Inc
 
Artificial neural network by arpit_sharma
Artificial neural network by arpit_sharmaArtificial neural network by arpit_sharma
Artificial neural network by arpit_sharma
Er. Arpit Sharma
 
Neuralnetwork 101222074552-phpapp02
Neuralnetwork 101222074552-phpapp02Neuralnetwork 101222074552-phpapp02
Neuralnetwork 101222074552-phpapp02
Deepu Gupta
 
ANN load forecasting
ANN load forecastingANN load forecasting
ANN load forecasting
Dr Ashok Tiwari
 
neuralnetwork.pptx
neuralnetwork.pptxneuralnetwork.pptx
neuralnetwork.pptx
SherinRappai1
 
neuralnetwork.pptx
neuralnetwork.pptxneuralnetwork.pptx
neuralnetwork.pptx
SherinRappai
 
Deep Learning Sample Class (Jon Lederman)
Deep Learning Sample Class (Jon Lederman)Deep Learning Sample Class (Jon Lederman)
Deep Learning Sample Class (Jon Lederman)
Jon Lederman
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
mustafa aadel
 
Artificial neural networks
Artificial neural networksArtificial neural networks
Artificial neural networks
madhu sudhakar
 
Neural network final NWU 4.3 Graphics Course
Neural network final NWU 4.3 Graphics CourseNeural network final NWU 4.3 Graphics Course
Neural network final NWU 4.3 Graphics Course
Mohaiminur Rahman
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
Mhd Khaled Alhalai
 
Neural-Networks.ppt
Neural-Networks.pptNeural-Networks.ppt
Neural-Networks.ppt
RINUSATHYAN
 
Neural
NeuralNeural
Neural
Vaibhav Shah
 

Similar to Introduction to Neural networks (under graduate course) Lecture 9 of 9 (20)

Unit 2 ml.pptx
Unit 2 ml.pptxUnit 2 ml.pptx
Unit 2 ml.pptx
 
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTS
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTSNEURAL NETWORK IN MACHINE LEARNING FOR STUDENTS
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTS
 
Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...
Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...
Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...
 
Artificial Neural Network (ANN
Artificial Neural Network (ANNArtificial Neural Network (ANN
Artificial Neural Network (ANN
 
Neural network
Neural networkNeural network
Neural network
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
 
Introduction to Perceptron and Neural Network.pptx
Introduction to Perceptron and Neural Network.pptxIntroduction to Perceptron and Neural Network.pptx
Introduction to Perceptron and Neural Network.pptx
 
Terminology Machine Learning
Terminology Machine LearningTerminology Machine Learning
Terminology Machine Learning
 
Artificial neural network by arpit_sharma
Artificial neural network by arpit_sharmaArtificial neural network by arpit_sharma
Artificial neural network by arpit_sharma
 
Neuralnetwork 101222074552-phpapp02
Neuralnetwork 101222074552-phpapp02Neuralnetwork 101222074552-phpapp02
Neuralnetwork 101222074552-phpapp02
 
ANN load forecasting
ANN load forecastingANN load forecasting
ANN load forecasting
 
neuralnetwork.pptx
neuralnetwork.pptxneuralnetwork.pptx
neuralnetwork.pptx
 
neuralnetwork.pptx
neuralnetwork.pptxneuralnetwork.pptx
neuralnetwork.pptx
 
Deep Learning Sample Class (Jon Lederman)
Deep Learning Sample Class (Jon Lederman)Deep Learning Sample Class (Jon Lederman)
Deep Learning Sample Class (Jon Lederman)
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
 
Artificial neural networks
Artificial neural networksArtificial neural networks
Artificial neural networks
 
Neural network final NWU 4.3 Graphics Course
Neural network final NWU 4.3 Graphics CourseNeural network final NWU 4.3 Graphics Course
Neural network final NWU 4.3 Graphics Course
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
 
Neural-Networks.ppt
Neural-Networks.pptNeural-Networks.ppt
Neural-Networks.ppt
 
Neural
NeuralNeural
Neural
 

More from Randa Elanwar

الجزء السادس ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء السادس ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوةالجزء السادس ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء السادس ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
Randa Elanwar
 
الجزء الخامس ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الخامس ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوةالجزء الخامس ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الخامس ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوةRanda Elanwar
 
الجزء الرابع ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الرابع ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوةالجزء الرابع ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الرابع ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
Randa Elanwar
 
الجزء الثالث ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الثالث ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوةالجزء الثالث ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الثالث ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوةRanda Elanwar
 
الجزء الثاني ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الثاني ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوةالجزء الثاني ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الثاني ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
Randa Elanwar
 
الجزء الأول ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الأول ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوةالجزء الأول ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الأول ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوةRanda Elanwar
 
تدريب مدونة علماء مصر على الكتابة الفنية (الترجمة والتلخيص )_Pdf5of5
تدريب مدونة علماء مصر على الكتابة الفنية (الترجمة والتلخيص    )_Pdf5of5تدريب مدونة علماء مصر على الكتابة الفنية (الترجمة والتلخيص    )_Pdf5of5
تدريب مدونة علماء مصر على الكتابة الفنية (الترجمة والتلخيص )_Pdf5of5
Randa Elanwar
 
تدريب مدونة علماء مصر على الكتابة الفنية (القصة القصيرة والخاطرة والأخطاء ال...
تدريب مدونة علماء مصر على الكتابة الفنية (القصة القصيرة والخاطرة  والأخطاء ال...تدريب مدونة علماء مصر على الكتابة الفنية (القصة القصيرة والخاطرة  والأخطاء ال...
تدريب مدونة علماء مصر على الكتابة الفنية (القصة القصيرة والخاطرة والأخطاء ال...
Randa Elanwar
 
تدريب مدونة علماء مصر على الكتابة الفنية (مقالات الموارد )_Pdf3of5
تدريب مدونة علماء مصر على الكتابة الفنية (مقالات الموارد   )_Pdf3of5تدريب مدونة علماء مصر على الكتابة الفنية (مقالات الموارد   )_Pdf3of5
تدريب مدونة علماء مصر على الكتابة الفنية (مقالات الموارد )_Pdf3of5
Randa Elanwar
 
تدريب مدونة علماء مصر على الكتابة الفنية (المقالات الإخبارية )_Pdf2of5
تدريب مدونة علماء مصر على الكتابة الفنية (المقالات الإخبارية  )_Pdf2of5تدريب مدونة علماء مصر على الكتابة الفنية (المقالات الإخبارية  )_Pdf2of5
تدريب مدونة علماء مصر على الكتابة الفنية (المقالات الإخبارية )_Pdf2of5
Randa Elanwar
 
تدريب مدونة علماء مصر على الكتابة الفنية (المقالات المبنية على البحث )_Pdf1of5
تدريب مدونة علماء مصر على الكتابة الفنية (المقالات المبنية على البحث )_Pdf1of5تدريب مدونة علماء مصر على الكتابة الفنية (المقالات المبنية على البحث )_Pdf1of5
تدريب مدونة علماء مصر على الكتابة الفنية (المقالات المبنية على البحث )_Pdf1of5
Randa Elanwar
 
تعريف بمدونة علماء مصر ومحاور التدريب على الكتابة للمدونين
تعريف بمدونة علماء مصر ومحاور التدريب على الكتابة للمدونينتعريف بمدونة علماء مصر ومحاور التدريب على الكتابة للمدونين
تعريف بمدونة علماء مصر ومحاور التدريب على الكتابة للمدونين
Randa Elanwar
 
Entrepreneurship_who_is_your_customer_(arabic)_7of7
Entrepreneurship_who_is_your_customer_(arabic)_7of7Entrepreneurship_who_is_your_customer_(arabic)_7of7
Entrepreneurship_who_is_your_customer_(arabic)_7of7
Randa Elanwar
 
Entrepreneurship_who_is_your_customer_(arabic)_5of7
Entrepreneurship_who_is_your_customer_(arabic)_5of7Entrepreneurship_who_is_your_customer_(arabic)_5of7
Entrepreneurship_who_is_your_customer_(arabic)_5of7
Randa Elanwar
 
Entrepreneurship_who_is_your_customer_(arabic)_4of7
Entrepreneurship_who_is_your_customer_(arabic)_4of7Entrepreneurship_who_is_your_customer_(arabic)_4of7
Entrepreneurship_who_is_your_customer_(arabic)_4of7
Randa Elanwar
 
Entrepreneurship_who_is_your_customer_(arabic)_2of7
Entrepreneurship_who_is_your_customer_(arabic)_2of7Entrepreneurship_who_is_your_customer_(arabic)_2of7
Entrepreneurship_who_is_your_customer_(arabic)_2of7
Randa Elanwar
 
يوميات طالب بدرجة مشرف (Part 19 of 20)
يوميات طالب بدرجة مشرف (Part 19 of 20)يوميات طالب بدرجة مشرف (Part 19 of 20)
يوميات طالب بدرجة مشرف (Part 19 of 20)
Randa Elanwar
 
يوميات طالب بدرجة مشرف (Part 18 of 20)
يوميات طالب بدرجة مشرف (Part 18 of 20)يوميات طالب بدرجة مشرف (Part 18 of 20)
يوميات طالب بدرجة مشرف (Part 18 of 20)Randa Elanwar
 
يوميات طالب بدرجة مشرف (Part 17 of 20)
يوميات طالب بدرجة مشرف (Part 17 of 20)يوميات طالب بدرجة مشرف (Part 17 of 20)
يوميات طالب بدرجة مشرف (Part 17 of 20)
Randa Elanwar
 
يوميات طالب بدرجة مشرف (Part 16 of 20)
يوميات طالب بدرجة مشرف (Part 16 of 20)يوميات طالب بدرجة مشرف (Part 16 of 20)
يوميات طالب بدرجة مشرف (Part 16 of 20)Randa Elanwar
 

More from Randa Elanwar (20)

الجزء السادس ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء السادس ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوةالجزء السادس ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء السادس ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
 
الجزء الخامس ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الخامس ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوةالجزء الخامس ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الخامس ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
 
الجزء الرابع ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الرابع ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوةالجزء الرابع ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الرابع ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
 
الجزء الثالث ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الثالث ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوةالجزء الثالث ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الثالث ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
 
الجزء الثاني ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الثاني ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوةالجزء الثاني ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الثاني ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
 
الجزء الأول ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الأول ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوةالجزء الأول ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
الجزء الأول ماذا ستقدم لعميلك ريادة الأعمال خطوة بخطوة
 
تدريب مدونة علماء مصر على الكتابة الفنية (الترجمة والتلخيص )_Pdf5of5
تدريب مدونة علماء مصر على الكتابة الفنية (الترجمة والتلخيص    )_Pdf5of5تدريب مدونة علماء مصر على الكتابة الفنية (الترجمة والتلخيص    )_Pdf5of5
تدريب مدونة علماء مصر على الكتابة الفنية (الترجمة والتلخيص )_Pdf5of5
 
تدريب مدونة علماء مصر على الكتابة الفنية (القصة القصيرة والخاطرة والأخطاء ال...
تدريب مدونة علماء مصر على الكتابة الفنية (القصة القصيرة والخاطرة  والأخطاء ال...تدريب مدونة علماء مصر على الكتابة الفنية (القصة القصيرة والخاطرة  والأخطاء ال...
تدريب مدونة علماء مصر على الكتابة الفنية (القصة القصيرة والخاطرة والأخطاء ال...
 
تدريب مدونة علماء مصر على الكتابة الفنية (مقالات الموارد )_Pdf3of5
تدريب مدونة علماء مصر على الكتابة الفنية (مقالات الموارد   )_Pdf3of5تدريب مدونة علماء مصر على الكتابة الفنية (مقالات الموارد   )_Pdf3of5
تدريب مدونة علماء مصر على الكتابة الفنية (مقالات الموارد )_Pdf3of5
 
تدريب مدونة علماء مصر على الكتابة الفنية (المقالات الإخبارية )_Pdf2of5
تدريب مدونة علماء مصر على الكتابة الفنية (المقالات الإخبارية  )_Pdf2of5تدريب مدونة علماء مصر على الكتابة الفنية (المقالات الإخبارية  )_Pdf2of5
تدريب مدونة علماء مصر على الكتابة الفنية (المقالات الإخبارية )_Pdf2of5
 
تدريب مدونة علماء مصر على الكتابة الفنية (المقالات المبنية على البحث )_Pdf1of5
تدريب مدونة علماء مصر على الكتابة الفنية (المقالات المبنية على البحث )_Pdf1of5تدريب مدونة علماء مصر على الكتابة الفنية (المقالات المبنية على البحث )_Pdf1of5
تدريب مدونة علماء مصر على الكتابة الفنية (المقالات المبنية على البحث )_Pdf1of5
 
تعريف بمدونة علماء مصر ومحاور التدريب على الكتابة للمدونين
تعريف بمدونة علماء مصر ومحاور التدريب على الكتابة للمدونينتعريف بمدونة علماء مصر ومحاور التدريب على الكتابة للمدونين
تعريف بمدونة علماء مصر ومحاور التدريب على الكتابة للمدونين
 
Entrepreneurship_who_is_your_customer_(arabic)_7of7
Entrepreneurship_who_is_your_customer_(arabic)_7of7Entrepreneurship_who_is_your_customer_(arabic)_7of7
Entrepreneurship_who_is_your_customer_(arabic)_7of7
 
Entrepreneurship_who_is_your_customer_(arabic)_5of7
Entrepreneurship_who_is_your_customer_(arabic)_5of7Entrepreneurship_who_is_your_customer_(arabic)_5of7
Entrepreneurship_who_is_your_customer_(arabic)_5of7
 
Entrepreneurship_who_is_your_customer_(arabic)_4of7
Entrepreneurship_who_is_your_customer_(arabic)_4of7Entrepreneurship_who_is_your_customer_(arabic)_4of7
Entrepreneurship_who_is_your_customer_(arabic)_4of7
 
Entrepreneurship_who_is_your_customer_(arabic)_2of7
Entrepreneurship_who_is_your_customer_(arabic)_2of7Entrepreneurship_who_is_your_customer_(arabic)_2of7
Entrepreneurship_who_is_your_customer_(arabic)_2of7
 
يوميات طالب بدرجة مشرف (Part 19 of 20)
يوميات طالب بدرجة مشرف (Part 19 of 20)يوميات طالب بدرجة مشرف (Part 19 of 20)
يوميات طالب بدرجة مشرف (Part 19 of 20)
 
يوميات طالب بدرجة مشرف (Part 18 of 20)
يوميات طالب بدرجة مشرف (Part 18 of 20)يوميات طالب بدرجة مشرف (Part 18 of 20)
يوميات طالب بدرجة مشرف (Part 18 of 20)
 
يوميات طالب بدرجة مشرف (Part 17 of 20)
يوميات طالب بدرجة مشرف (Part 17 of 20)يوميات طالب بدرجة مشرف (Part 17 of 20)
يوميات طالب بدرجة مشرف (Part 17 of 20)
 
يوميات طالب بدرجة مشرف (Part 16 of 20)
يوميات طالب بدرجة مشرف (Part 16 of 20)يوميات طالب بدرجة مشرف (Part 16 of 20)
يوميات طالب بدرجة مشرف (Part 16 of 20)
 

Recently uploaded

IoT (Internet of Things) introduction Notes.pdf
IoT (Internet of Things) introduction Notes.pdfIoT (Internet of Things) introduction Notes.pdf
IoT (Internet of Things) introduction Notes.pdf
roshanranjit222
 
Interprofessional Education Platform Introduction.pdf
Interprofessional Education Platform Introduction.pdfInterprofessional Education Platform Introduction.pdf
Interprofessional Education Platform Introduction.pdf
Ben Aldrich
 
The Rise of the Digital Telecommunication Marketplace.pptx
The Rise of the Digital Telecommunication Marketplace.pptxThe Rise of the Digital Telecommunication Marketplace.pptx
The Rise of the Digital Telecommunication Marketplace.pptx
PriyaKumari928991
 
Keynote given on June 24 for MASSP at Grand Traverse City
Keynote given on June 24 for MASSP at Grand Traverse CityKeynote given on June 24 for MASSP at Grand Traverse City
Keynote given on June 24 for MASSP at Grand Traverse City
PJ Caposey
 
Environmental science 1.What is environmental science and components of envir...
Environmental science 1.What is environmental science and components of envir...Environmental science 1.What is environmental science and components of envir...
Environmental science 1.What is environmental science and components of envir...
Deepika
 
pol sci Election and Representation Class 11 Notes.pdf
pol sci Election and Representation Class 11 Notes.pdfpol sci Election and Representation Class 11 Notes.pdf
pol sci Election and Representation Class 11 Notes.pdf
BiplabHalder13
 
How to Create a Stage or a Pipeline in Odoo 17 CRM
How to Create a Stage or a Pipeline in Odoo 17 CRMHow to Create a Stage or a Pipeline in Odoo 17 CRM
How to Create a Stage or a Pipeline in Odoo 17 CRM
Celine George
 
Post init hook in the odoo 17 ERP Module
Post init hook in the  odoo 17 ERP ModulePost init hook in the  odoo 17 ERP Module
Post init hook in the odoo 17 ERP Module
Celine George
 
India Quiz (Prelims and Finals) by Quiz Club, IIT Kanpur
India Quiz (Prelims and Finals) by Quiz Club, IIT KanpurIndia Quiz (Prelims and Finals) by Quiz Club, IIT Kanpur
India Quiz (Prelims and Finals) by Quiz Club, IIT Kanpur
Quiz Club IIT Kanpur
 
Information and Communication Technology in Education
Information and Communication Technology in EducationInformation and Communication Technology in Education
Information and Communication Technology in Education
MJDuyan
 
Science-9-Lesson-1-The Bohr Model-NLC.pptx pptx
Science-9-Lesson-1-The Bohr Model-NLC.pptx pptxScience-9-Lesson-1-The Bohr Model-NLC.pptx pptx
Science-9-Lesson-1-The Bohr Model-NLC.pptx pptx
Catherine Dela Cruz
 
220711130095 Tanu Pandey message currency, communication speed & control EPC ...
220711130095 Tanu Pandey message currency, communication speed & control EPC ...220711130095 Tanu Pandey message currency, communication speed & control EPC ...
220711130095 Tanu Pandey message currency, communication speed & control EPC ...
Kalna College
 
Decolonizing Universal Design for Learning
Decolonizing Universal Design for LearningDecolonizing Universal Design for Learning
Decolonizing Universal Design for Learning
Frederic Fovet
 
Brand Guideline of Bashundhara A4 Paper - 2024
Brand Guideline of Bashundhara A4 Paper - 2024Brand Guideline of Bashundhara A4 Paper - 2024
Brand Guideline of Bashundhara A4 Paper - 2024
khabri85
 
Talking Tech through Compelling Visual Aids
Talking Tech through Compelling Visual AidsTalking Tech through Compelling Visual Aids
Talking Tech through Compelling Visual Aids
MattVassar1
 
220711130086 Sukanta Singh E learning and mobile learning EPC 3 Internal Asse...
220711130086 Sukanta Singh E learning and mobile learning EPC 3 Internal Asse...220711130086 Sukanta Singh E learning and mobile learning EPC 3 Internal Asse...
220711130086 Sukanta Singh E learning and mobile learning EPC 3 Internal Asse...
Kalna College
 
220711130082 Srabanti Bag Internet Resources For Natural Science
220711130082 Srabanti Bag Internet Resources For Natural Science220711130082 Srabanti Bag Internet Resources For Natural Science
220711130082 Srabanti Bag Internet Resources For Natural Science
Kalna College
 
Erasmus + DISSEMINATION ACTIVITIES Croatia
Erasmus + DISSEMINATION ACTIVITIES CroatiaErasmus + DISSEMINATION ACTIVITIES Croatia
Erasmus + DISSEMINATION ACTIVITIES Croatia
whatchangedhowreflec
 
The Science of Learning: implications for modern teaching
The Science of Learning: implications for modern teachingThe Science of Learning: implications for modern teaching
The Science of Learning: implications for modern teaching
Derek Wenmoth
 
A Quiz on Drug Abuse Awareness by Quizzito
A Quiz on Drug Abuse Awareness by QuizzitoA Quiz on Drug Abuse Awareness by Quizzito
A Quiz on Drug Abuse Awareness by Quizzito
Quizzito The Quiz Society of Gargi College
 

Recently uploaded (20)

IoT (Internet of Things) introduction Notes.pdf
IoT (Internet of Things) introduction Notes.pdfIoT (Internet of Things) introduction Notes.pdf
IoT (Internet of Things) introduction Notes.pdf
 
Interprofessional Education Platform Introduction.pdf
Interprofessional Education Platform Introduction.pdfInterprofessional Education Platform Introduction.pdf
Interprofessional Education Platform Introduction.pdf
 
The Rise of the Digital Telecommunication Marketplace.pptx
The Rise of the Digital Telecommunication Marketplace.pptxThe Rise of the Digital Telecommunication Marketplace.pptx
The Rise of the Digital Telecommunication Marketplace.pptx
 
Keynote given on June 24 for MASSP at Grand Traverse City
Keynote given on June 24 for MASSP at Grand Traverse CityKeynote given on June 24 for MASSP at Grand Traverse City
Keynote given on June 24 for MASSP at Grand Traverse City
 
Environmental science 1.What is environmental science and components of envir...
Environmental science 1.What is environmental science and components of envir...Environmental science 1.What is environmental science and components of envir...
Environmental science 1.What is environmental science and components of envir...
 
pol sci Election and Representation Class 11 Notes.pdf
pol sci Election and Representation Class 11 Notes.pdfpol sci Election and Representation Class 11 Notes.pdf
pol sci Election and Representation Class 11 Notes.pdf
 
How to Create a Stage or a Pipeline in Odoo 17 CRM
How to Create a Stage or a Pipeline in Odoo 17 CRMHow to Create a Stage or a Pipeline in Odoo 17 CRM
How to Create a Stage or a Pipeline in Odoo 17 CRM
 
Post init hook in the odoo 17 ERP Module
Post init hook in the  odoo 17 ERP ModulePost init hook in the  odoo 17 ERP Module
Post init hook in the odoo 17 ERP Module
 
India Quiz (Prelims and Finals) by Quiz Club, IIT Kanpur
India Quiz (Prelims and Finals) by Quiz Club, IIT KanpurIndia Quiz (Prelims and Finals) by Quiz Club, IIT Kanpur
India Quiz (Prelims and Finals) by Quiz Club, IIT Kanpur
 
Information and Communication Technology in Education
Information and Communication Technology in EducationInformation and Communication Technology in Education
Information and Communication Technology in Education
 
Science-9-Lesson-1-The Bohr Model-NLC.pptx pptx
Science-9-Lesson-1-The Bohr Model-NLC.pptx pptxScience-9-Lesson-1-The Bohr Model-NLC.pptx pptx
Science-9-Lesson-1-The Bohr Model-NLC.pptx pptx
 
220711130095 Tanu Pandey message currency, communication speed & control EPC ...
220711130095 Tanu Pandey message currency, communication speed & control EPC ...220711130095 Tanu Pandey message currency, communication speed & control EPC ...
220711130095 Tanu Pandey message currency, communication speed & control EPC ...
 
Decolonizing Universal Design for Learning
Decolonizing Universal Design for LearningDecolonizing Universal Design for Learning
Decolonizing Universal Design for Learning
 
Brand Guideline of Bashundhara A4 Paper - 2024
Brand Guideline of Bashundhara A4 Paper - 2024Brand Guideline of Bashundhara A4 Paper - 2024
Brand Guideline of Bashundhara A4 Paper - 2024
 
Talking Tech through Compelling Visual Aids
Talking Tech through Compelling Visual AidsTalking Tech through Compelling Visual Aids
Talking Tech through Compelling Visual Aids
 
220711130086 Sukanta Singh E learning and mobile learning EPC 3 Internal Asse...
220711130086 Sukanta Singh E learning and mobile learning EPC 3 Internal Asse...220711130086 Sukanta Singh E learning and mobile learning EPC 3 Internal Asse...
220711130086 Sukanta Singh E learning and mobile learning EPC 3 Internal Asse...
 
220711130082 Srabanti Bag Internet Resources For Natural Science
220711130082 Srabanti Bag Internet Resources For Natural Science220711130082 Srabanti Bag Internet Resources For Natural Science
220711130082 Srabanti Bag Internet Resources For Natural Science
 
Erasmus + DISSEMINATION ACTIVITIES Croatia
Erasmus + DISSEMINATION ACTIVITIES CroatiaErasmus + DISSEMINATION ACTIVITIES Croatia
Erasmus + DISSEMINATION ACTIVITIES Croatia
 
The Science of Learning: implications for modern teaching
The Science of Learning: implications for modern teachingThe Science of Learning: implications for modern teaching
The Science of Learning: implications for modern teaching
 
A Quiz on Drug Abuse Awareness by Quizzito
A Quiz on Drug Abuse Awareness by QuizzitoA Quiz on Drug Abuse Awareness by Quizzito
A Quiz on Drug Abuse Awareness by Quizzito
 

Introduction to Neural networks (under graduate course) Lecture 9 of 9

  • 1. Neural Networks Dr. Randa Elanwar Lecture 9
  • 2. Lecture Content • Mapping networks: – Back-propagation neural network – Self-organizing map – Counter propagation network • Spatiotemporal Network • Stochastic Networks – Boltzmann machine • Neurocognition network 2Neural Networks Dr. Randa Elanwar
  • 3. Mapping networks • When the problem is non linear and no straight line could ever separate samples in the feature space we need multilayer perceptrons (having hidden layer’s’) to achieve nonlinearity. • The idea is that we map/transform/translate our data to another feature space that is linearly separable. Thus we call them mapping networks. • We will discuss three types of mapping networks: the back-propagation neural network, self-organizing map, counter propagation network. 3Neural Networks Dr. Randa Elanwar
  • 4. Mapping networks • Networks without hidden units are very limited in the input-output mappings they can model. – More layers of linear units do not help. Its still linear. – Fixed output non-linearities are not enough • We need multiple layers of adaptive non-linear hidden units. • But how can we train such nets? – We need an efficient way of adapting all the weights, not just the last layer. i.e., Learning the weights going into hidden units . This is hard. – Why? – Because: Nobody is telling us directly what hidden units should do. – Solution: This can be achieved using ‘Backpropagation’ learning 4Neural Networks Dr. Randa Elanwar
  • 5. Learning with hidden layers • Mathematically, the learning process is an optimization problem. We initiate the NN system with some parameters (weights) and use known examples to find out the optimal values of such weights. • Generally, the solution of an optimization problem is to find the parameter value that leads to minimum value of an optimization function. 5Neural Networks Dr. Randa Elanwar G(t) t In our case, the optimization function that we need to minimize to get the final weights is the error function E = ydes-yact E = ydes-f(W.X) To get the minimum value mathematically we differentiate the error function with respect to the parameter we need to get we call it W E  
  • 6. Learning with hidden layers • We define the “gradient”: w = . . X • If  is +ve this means that the current values of W makes the differentiation result +ve which is wrong. We want differentiation result to be = 0 (minimum point) we must move in the opposite direction of the gradient (subtract). The opposite is also true. • If  is =0 this means that the current values of W makes the differentiation result = 0 which is right. These weights are the optimal values (solution) and w should stop the algorithm. The network now is trained and ready for use. 6Neural Networks Dr. Randa Elanwar
  • 7. The back propagation algorithm • The backpropagation learning algorithm can be divided into two phases: propagation and weight update. 7Neural Networks Dr. Randa Elanwar Phase 1: Propagation 1.Forward propagation of a training pattern's input through the neural network in order to generate the propagation's output activations (yact). 2.Backward propagation of the propagation's output activations through the neural network using the training pattern's target (ydes) in order to generate the deltas () of all output and hidden neurons. Phase 2: Weight update For each weight follow the following steps: 1.Multiply its output delta () and input activation (x) and the learning rate () to get the gradient of the weight (w). 2.Bring the weight in the opposite direction of the gradient by subtracting it from the weight. - The sign of the gradient of a weight indicates where the error is increasing, this is why the weight must be updated in the opposite direction. - Repeat phase 1 and 2 until the performance of the network is satisfactory.
  • 8. Backpropagation Networks • They are the nonlinear (mapping) neural networks using the backpropagation supervised learning technique. • Modes of learning of nonlinear nets: • There are three modes of learning to choose from: on-line (pattern), batch and stochastic. • In on-line and stochastic learning, each propagation is followed immediately by a weight update. • In batch learning, many propagations occur before updating the weights. • Batch learning requires more memory capacity, but on-line and stochastic learning require more updates. 8Neural Networks Dr. Randa Elanwar
  • 9. Backpropagation Networks • On-line learning is used for dynamic environments that provide a continuous stream of new patterns. • Stochastic learning and batch learning both make use of a training set of static patterns. Stochastic goes through the data set in a random order in order to reduce its chances of getting stuck in local minima. • Stochastic learning is also much faster than batch learning since weights are updated immediately after each propagation. Yet batch learning will yield a much more stable descent to a local minima since each update is performed based on all patterns. 9Neural Networks Dr. Randa Elanwar
  • 10. Backpropagation Networks • Applications of supervised learning (Backpropagation NN) include • Pattern recognition • Credit approval • Target marketing • Medical diagnosis • Defective parts identification in manufacturing • Crime zoning • Treatment effectiveness analysis • Etc 10Neural Networks Dr. Randa Elanwar
  • 11. Self-organizing map • We can also train networks where there is no teacher. This is called unsupervised learning. The network learns a prototype based on the distribution of patterns in the training data. Such networks allow us to: – Discover underlying structure of the data – Encode or compress the data – Transform the data • Self-organizing maps (SOMs) are a data visualization technique invented by Professor Teuvo Kohonen – Also called Kohonen Networks, Competitive Learning, Winner-Take-All Learning – Generally reduces the dimensions of data through the use of self- organizing neural networks – Useful for data visualization; humans cannot visualize high dimensional data so this is often a useful technique to make sense of large data sets 11Neural Networks Dr. Randa Elanwar
  • 12. Self-organizing map • SOM structure: 1. Weights in neuron must represent a class of pattern. We have a neuron for each class. 2. Inputs pattern presented to all neurons and each produces an output. Output: measure of the match between input pattern and pattern stored by neuron. 3. A competitive learning strategy selects neuron with largest response. 4. A method of reinforcing the largest response. 12Neural Networks Dr. Randa Elanwar
  • 13. Self-organizing map • Unsupervised classification learning is based on clustering of input data. No a priori knowledge is about an input’s membership in a particular class. • Instead, gradually detected characteristics and a history of training will be used to assist the network in defining classes and possible boundaries between them. • Clustering is understood to be the grouping of similar objects and separating of dissimilar ones. • We discuss Kohonen’s network which classifies input vectors into one of the specified number of m categories, according to the clusters detected in the training set 13Neural Networks Dr. Randa Elanwar
  • 14. Kohonen’s Network 14Neural Networks Dr. Randa Elanwar Kohonen network X •The Kohonen network is a self-organising network with the following characteristics: 1. Neurons are arranged on a 2D grid 2. Inputs are sent to all neurons 3. There are no connections between neurons 4. For a neuron output (j) is a weighted sum of multiplication of x and w vectors, where x is the input, w is the weights 5. There is no threshold or bias 6. Input values and weights are normalized
  • 15. Self-organizing map Learning in Kohonen networks: • Initially the weights in each neuron are random • Input values are sent to all the neurons • The outputs of each neuron are compared • The “winner” is the neuron with the largest output value • Having found the winner, the weights of the winning neuron are adjusted • Weights of neurons in a surrounding neighbourhood are also adjusted • As training progresses the neighbourhood gets smaller • Weights are adjusted according to the following formula: 15Neural Networks Dr. Randa Elanwar
  • 16. Self-organizing map • The learning coefficient (alpha) starts with a value of 1 and gradually reduces to 0 • This has the effect of making big changes to the weights initially, but no changes at the end • The weights are adjusted so that they more closely resemble the input patterns Applications of unsupervised learning (Kohonen’s NN) include • Clustering • Vector quantization • Data compression • Feature extraction 16Neural Networks Dr. Randa Elanwar
  • 17. Counter propagation network • The counterpropagation network (CPN) is a fast-learning combination of unsupervised and supervised learning. • Although this network uses linear neurons, it can learn nonlinear functions by means of a hidden layer of competitive units. • Moreover, the network is able to learn a function and its inverse at the same time. • However, to simplify things, we will only consider the feedforward mechanism of the CPN. 17Neural Networks Dr. Randa Elanwar
  • 18. Counter propagation network • Training: 1.Randomly select a vector pair (x, y) from the training set. 2.Measure the similarity between the input vector and the activation of the hidden-layer units. 3.In the hidden (competitive) layer, determine the unit with the largest activation (the winner). I.e., the neuron whose weight vector is most similar to the current input vector is the “winner.” 4.Adjust the connection weights inbetween 5.Repeat until each input pattern is consistently associated with the same competitive unit. 18Neural Networks Dr. Randa Elanwar
  • 19. Counter propagation network • After the first phase of the training, each hidden-layer neuron is associated with a subset of input vectors (class of patterns). • In the second phase of the training, we adjust the weights in the network’s output layer in such a way that, for any winning hidden- layer unit, the network’s output is as close as possible to the desired output for the winning unit’s associated input vectors. • The idea is that when we later use the network to compute functions, the output of the winning hidden-layer unit is 1, and the output of all other hidden-layer units is 0. 19Neural Networks Dr. Randa Elanwar
  • 20. Spatiotemporal Networks •A spatio-temporal neural net differs from other neural networks in two ways: 1. Neurons has recurrent links that have different propagation delays 2. The state of the network depends not only on which nodes are firing, but also on the relative firing times of nodes. i.e., the significance of a node varies with time and depends on the firing state of other nodes. •The use of recurrence and multiple links with variable propagation delays provides a rich mechanism for feature extraction and pattern recognition: 1. Recurrent links enable nodes to integrate and differentiate inputs. I.e., detect features 2. multiple links with variable propagation delays between nodes serve as a short-term memory. 20Neural Networks Dr. Randa Elanwar
  • 21. Spatiotemporal Networks • Applications: • Problems such as speech recognition and time series prediction where the input signal has an explicit temporal aspect. • Tasks like image recognition do not have an explicit temporal aspect, but can also be done by converting static patterns into time-varying (spatio- temporal) signals via scanning the image. This would lead to a number of significant advantages: – The recognition system becomes ‘shift invarient’ – The spatio-temporal approach explains the image geometry since the local spatial relationships in the image are expressed as local temporal variations in the scanned input. – Reduction of complexity (from 2D to 1D) – The scanning approach allows a visual pattern recognition system to deal with inputs of arbitrary extent (not only static fixed 2D pattern) 21Neural Networks Dr. Randa Elanwar
  • 22. Stochastic neural networks • Stochastic neural networks are a type of artificial neural networks, which is a tool of artificial intelligence. They are built by introducing random variations into the network, either by giving the network's neurons stochastic transfer functions, or by giving them stochastic weights. This makes them useful tools for optimization problems, since the random fluctuations help it escape from local minima. • Stochastic neural networks that are built by using stochastic transfer functions are often called Boltzmann machines. • Stochastic neural networks have found applications in risk management, oncology, bioinformatics, and other similar fields 22Neural Networks Dr. Randa Elanwar
  • 23. Stochastic Networks: Boltzmann machine • The neurons are stochastic: at any time there is a probability attached to whether the neurons fires. • Used for solving constrained optimization problems. • Typical Boltzmann Machine: – Weights are fixed to represent the constrains of the problem and the function to be optimized. – The net seeks the solution by changing the activations of the units (0 or 1) based on a probability distribution and the effect that the change would have on the energy function or consensus function for the net. • May use either supervised or unsupervised learning. • Learning in Boltzmann Machine is accomplished by using a Simulated Annealing technique which has stochastic nature. This is used to reduce the probability of the net becoming trapped in a local minimum which is not a global minimum. 23Neural Networks Dr. Randa Elanwar
  • 24. Stochastic Networks: Boltzmann machine • Learning characteristics: – Each neuron fires with bipolar values. – All connections are symmetric. – In activation passing, the next neuron whose state we wish to update is selected randomly. – There are no self-feedback (connections from a neuron to itself) 24Neural Networks Dr. Randa Elanwar
  • 25. Stochastic Networks: Boltzmann machine • There are three phases in operation of the network: – The clamped phase in which the input and output of visible neurons are held fixed, while the hidden neurons are allowed to vary. – The free running phase in which only the inputs are held fixed and other neurons are allowed to vary. – The learning phase. • These phases iterate till learning has created a Boltzmann Machine which can be said to have learned the input patterns and will converge to the learned patterns when noisy or incomplete pattern is presented. 25Neural Networks Dr. Randa Elanwar
  • 26. Stochastic Networks: Boltzmann machine • For unsupervised learning Generally the initial weights of the net are randomly set to values in a small range e.g. -0.5 to +0.5. • Then an input pattern is presented to the net and clamped to the visible neurons. • choose a hidden neurons at random and flip its state from sj to –sj according to certain probability distribution • The activation passing can continue till the net hidden neurons reach equilibrium. • During free running phase, after presentation of the input patterns all neurons can update their states. • The learning phase depends whether weight are changed depend on the difference between the "real" distribution (neuron state) in clamped phase and the one which will be produced (eventually) by the machine in free mode. 26Neural Networks Dr. Randa Elanwar
  • 27. Stochastic Networks: Boltzmann machine • For supervised learning the set of visible neurons is split into input and output neurons, and the machine will be used to associate an input pattern with an output pattern. • During the clamped phase, the input and output patterns are clamped to the appropriate units. • The hidden neurons’ activations can settle at the various values. • During free running phase, only the input neurons are clamped – both the output neurons and the hidden neurons can pass activation round till the activations in the network settles. • Learning rule here is the same as before but must be modulated (multiplied) by the probability of the input’s patterns 27Neural Networks Dr. Randa Elanwar
  • 28. Neurocognition network • Neurocognitive networks are large-scale systems of distributed and interconnected neuronal populations in the central nervous system organized to perform cognitive functions. • many computer scientists try to simulate human cognition with computers. This line of research can be roughly split into two types: research seeking to create machines as adept as humans (or more so), and research attempting to figure out the computational basis of human cognition — that is, how the brain actually carries out its computations. This latter branch of research can be called computational modeling (while the former is often called artificial intelligence or AI). 28Neural Networks Dr. Randa Elanwar
  翻译: