尊敬的 微信汇率:1円 ≈ 0.046166 元 支付宝汇率:1円 ≈ 0.046257元 [退出登录]
SlideShare a Scribd company logo
Next Assignment ,[object Object],[object Object],[object Object],[object Object]
ART1 Demo Increasing  vigilance  causes the network to be more selective, to introduce a new prototype when the fit is not good. Try different patterns
Hebbian Learning
Hebb’s Postulate “ When an axon of cell A is near enough to excite a cell B and repeatedly or persistently takes part in firing it, some growth process or metabolic change takes place in one or both cells such that A’s efficiency, as one of the cells firing B, is increased.” D. O. Hebb, 1949 A B In other words, when a weight contributes to firing a neuron, the weight is increased. (If the neuron doesn’t fire, then it is not).
A B
A B
Colloquial Corollaries ,[object Object]
Colloquial Corollaries? ,[object Object]
Generalized Hebb Rule ,[object Object],[object Object]
Flavors of Hebbian Learning ,[object Object],[object Object],[object Object],[object Object]
Unsupervised Hebbian Learning (aka Associative Learning)
Simple Associative Network input output
Banana Associator Unconditioned Stimulus Conditioned Stimulus Didn’t Pavlov anticipate this?
Banana Associator Demo can be toggled
Unsupervised Hebb Rule Vector Form: Training Sequence: actual response input
Learning Banana Smell Initial Weights: Training Sequence: First Iteration (sight fails, smell present):    = 1 unconditioned (shape) conditioned (smell) a 1   h a r d l i m w 0 p 0 1   w 0   p 1   0.5 – +   h a r d l i m 1 0  0 1  0.5 – +   0 (no banana) = = =
Example Second Iteration (sight works, smell present): Third Iteration (sight fails, smell present): Banana will now be detected if either sensor works. a 2   h a r d l i m w 0 p 0 2   w 1   p 2   0.5 – +   h a r d l i m 1 1  0 1  0.5 – +   1 (banana) = = =
Problems with Hebb Rule ,[object Object],[object Object]
Hebb Rule with Decay This keeps the weight matrix from growing without bound, which can be demonstrated by setting both  a i  and  p j  to 1:
Banana Associator with Decay
Example: Banana Associator First Iteration (sight fails, smell present): Second Iteration (sight works, smell present):    = 0.1    = 1 a 1   h a r d l i m w 0 p 0 1   w 0   p 1   0.5 – +   h a r d l i m 1 0  0 1  0.5 – +   0 (no banana) = = = a 2   h a r d l i m w 0 p 0 2   w 1   p 2   0.5 – +   h a r d l i m 1 1  0 1  0.5 – +   1 (banana) = = =
Example Third Iteration (sight fails, smell present):
General Decay Demo no decay larger decay w i j m a x   - - - =
Problem of Hebb with Decay Associations  will be lost if stimuli are not occasionally presented. If  a i  = 0, then If    = 0, this becomes Therefore the weight decays by 10% at each iteration where there is no stimulus.
Solution to Hebb Decay Problem ,[object Object],[object Object]
Instar (Recognition Network)
Instar Operation The instar will be active when or For normalized vectors, the largest inner product occurs when the angle between the weight vector and the input vector is zero -- the input vector is equal to the weight vector. The rows of a weight matrix represent patterns to be recognized. w T 1 p w 1 p  cos b –  =
Vector Recognition If we set the instar will only be active when    =  0. If we set the instar will be active for a range of angles.  As  b  is increased, the more patterns there will be (over a wider range of   ) which will activate the instar. b w 1 p – = b w 1 p – > w 1
Instar Rule Hebb with Decay Modify so that  learning and forgetting will only occur when the neuron is active  - Instar Rule: or Vector Form: w i j q   w i j q 1 –    a i q   p j q    a i q   w q 1 –   – + = i j
Graphical Representation For the case where the instar is active ( a i  =   1): or For the case where the instar is inactive ( a i  =   0):
Instar Demo weight vector input vector  W
Outstar ( Recall  Network)
Outstar Operation Suppose we want the outstar to recall a certain pattern  a *  whenever the input  p   =   1 is presented to the network. Let  Then, when  p   =   1 and the pattern is correctly recalled. The columns of a weight matrix represent patterns  to be recalled.
Outstar Rule For the instar rule we made the weight decay term of the Hebb rule proportional to the  output  of the network.  For the outstar rule we make the weight decay term proportional to the  input  of the network. If we make the decay rate    equal to the learning rate   , Vector Form:
Example - Pineapple Recall
Definitions
Outstar Demo
Iteration 1    = 1
Convergence
Supervised Hebbian Learning
Linear Associator Training Set:
Hebb Rule Presynaptic Signal Postsynaptic Signal Simplified Form: Supervised Form: Matrix Form: actual output input pattern desired  output
Batch Operation Matrix Form: (Zero Initial Weights)  W t 1 t 2  t Q p 1 T p 2 T p Q T T P T = = T t 1 t 2  t Q = P p 1 p 2  p Q =
Performance Analysis Case I, input patterns are orthogonal. Therefore the network output equals the target: Case II, input patterns are normalized, but not orthogonal. Error term 0 q k  =
Example Banana Apple Normalized Prototype Patterns Weight Matrix (Hebb Rule): Tests: Banana Apple
Pseudoinverse Rule - (1) Performance Index: Matrix Form: Mean-squared error T t 1 t 2  t Q = P p 1 p 2  p Q = || E || 2 e i j 2 j  i  =
Pseudoinverse Rule - (2) Minimize: If an inverse exists for  P ,  F ( W ) can be made zero: When an inverse does not exist   F ( W ) can be minimized using the pseudoinverse:
Relationship to the Hebb Rule Hebb Rule Pseudoinverse Rule If the prototype patterns are orthonormal: W T P T =
Example
Autoassociative Memory
Tests 50% Occluded 67% Occluded Noisy Patterns (7 pixels)
Supervised Hebbian Demo
Spectrum of Hebbian Learning Basic Supervised Rule: Supervised with Learning Rate: Smoothing: Delta Rule: Unsupervised: target actual

More Related Content

What's hot

Logics for non monotonic reasoning-ai
Logics for non monotonic reasoning-aiLogics for non monotonic reasoning-ai
Logics for non monotonic reasoning-ai
ShaishavShah8
 
Feed forward ,back propagation,gradient descent
Feed forward ,back propagation,gradient descentFeed forward ,back propagation,gradient descent
Feed forward ,back propagation,gradient descent
Muhammad Rasel
 
Associative memory network
Associative memory networkAssociative memory network
Associative memory network
Dr. C.V. Suresh Babu
 
Adaptive Resonance Theory
Adaptive Resonance TheoryAdaptive Resonance Theory
Adaptive Resonance Theory
Naveen Kumar
 
Counter propagation Network
Counter propagation NetworkCounter propagation Network
Counter propagation Network
Akshay Dhole
 
Control Strategies in AI
Control Strategies in AIControl Strategies in AI
Control Strategies in AI
Amey Kerkar
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
Prakash K
 
Radial basis function network ppt bySheetal,Samreen and Dhanashri
Radial basis function network ppt bySheetal,Samreen and DhanashriRadial basis function network ppt bySheetal,Samreen and Dhanashri
Radial basis function network ppt bySheetal,Samreen and Dhanashri
sheetal katkar
 
Uncertain Knowledge and Reasoning in Artificial Intelligence
Uncertain Knowledge and Reasoning in Artificial IntelligenceUncertain Knowledge and Reasoning in Artificial Intelligence
Uncertain Knowledge and Reasoning in Artificial Intelligence
Experfy
 
Machine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural NetworksMachine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural Networks
Francesco Collova'
 
Optimization in Deep Learning
Optimization in Deep LearningOptimization in Deep Learning
Optimization in Deep Learning
Yan Xu
 
Feedforward neural network
Feedforward neural networkFeedforward neural network
Feedforward neural network
Sopheaktra YONG
 
Unit3 dip
Unit3 dipUnit3 dip
Unit3 dip
Imran Khan
 
Back propagation
Back propagationBack propagation
Back propagation
Nagarajan
 
Neural networks...
Neural networks...Neural networks...
Neural networks...
Molly Chugh
 
Optimization for Deep Learning
Optimization for Deep LearningOptimization for Deep Learning
Optimization for Deep Learning
Sebastian Ruder
 
Heuristc Search Techniques
Heuristc Search TechniquesHeuristc Search Techniques
Heuristc Search Techniques
Jismy .K.Jose
 
Probabilistic Reasoning
Probabilistic ReasoningProbabilistic Reasoning
Probabilistic Reasoning
Junya Tanaka
 
Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)
Gaurav Mittal
 
Activation function
Activation functionActivation function
Activation function
Astha Jain
 

What's hot (20)

Logics for non monotonic reasoning-ai
Logics for non monotonic reasoning-aiLogics for non monotonic reasoning-ai
Logics for non monotonic reasoning-ai
 
Feed forward ,back propagation,gradient descent
Feed forward ,back propagation,gradient descentFeed forward ,back propagation,gradient descent
Feed forward ,back propagation,gradient descent
 
Associative memory network
Associative memory networkAssociative memory network
Associative memory network
 
Adaptive Resonance Theory
Adaptive Resonance TheoryAdaptive Resonance Theory
Adaptive Resonance Theory
 
Counter propagation Network
Counter propagation NetworkCounter propagation Network
Counter propagation Network
 
Control Strategies in AI
Control Strategies in AIControl Strategies in AI
Control Strategies in AI
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
 
Radial basis function network ppt bySheetal,Samreen and Dhanashri
Radial basis function network ppt bySheetal,Samreen and DhanashriRadial basis function network ppt bySheetal,Samreen and Dhanashri
Radial basis function network ppt bySheetal,Samreen and Dhanashri
 
Uncertain Knowledge and Reasoning in Artificial Intelligence
Uncertain Knowledge and Reasoning in Artificial IntelligenceUncertain Knowledge and Reasoning in Artificial Intelligence
Uncertain Knowledge and Reasoning in Artificial Intelligence
 
Machine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural NetworksMachine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural Networks
 
Optimization in Deep Learning
Optimization in Deep LearningOptimization in Deep Learning
Optimization in Deep Learning
 
Feedforward neural network
Feedforward neural networkFeedforward neural network
Feedforward neural network
 
Unit3 dip
Unit3 dipUnit3 dip
Unit3 dip
 
Back propagation
Back propagationBack propagation
Back propagation
 
Neural networks...
Neural networks...Neural networks...
Neural networks...
 
Optimization for Deep Learning
Optimization for Deep LearningOptimization for Deep Learning
Optimization for Deep Learning
 
Heuristc Search Techniques
Heuristc Search TechniquesHeuristc Search Techniques
Heuristc Search Techniques
 
Probabilistic Reasoning
Probabilistic ReasoningProbabilistic Reasoning
Probabilistic Reasoning
 
Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)
 
Activation function
Activation functionActivation function
Activation function
 

Similar to Hebbian Learning

Neural Networks
Neural NetworksNeural Networks
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
ssuserab4f3e
 
Unit 2
Unit 2Unit 2
constructing_generic_algorithms__ben_deane__cppcon_2020.pdf
constructing_generic_algorithms__ben_deane__cppcon_2020.pdfconstructing_generic_algorithms__ben_deane__cppcon_2020.pdf
constructing_generic_algorithms__ben_deane__cppcon_2020.pdf
SayanSamanta39
 
Associative_Memory_Neural_Networks_pptx.pptx
Associative_Memory_Neural_Networks_pptx.pptxAssociative_Memory_Neural_Networks_pptx.pptx
Associative_Memory_Neural_Networks_pptx.pptx
dgfsdf1
 
Lec10
Lec10Lec10
CS532L4_Backpropagation.pptx
CS532L4_Backpropagation.pptxCS532L4_Backpropagation.pptx
CS532L4_Backpropagation.pptx
MFaisalRiaz5
 
NN-Ch3 (1).ppt
NN-Ch3 (1).pptNN-Ch3 (1).ppt
NN-Ch3 (1).ppt
RafeeqAhmed42
 
CS767_Lecture_04.pptx
CS767_Lecture_04.pptxCS767_Lecture_04.pptx
CS767_Lecture_04.pptx
ShujatHussainGadi
 
Unit iii update
Unit iii updateUnit iii update
Unit iii update
Indira Priyadarsini
 
S1 Dualsimplex
S1 DualsimplexS1 Dualsimplex
S1 Dualsimplex
Ngo Hung Long
 
Perceptron
PerceptronPerceptron
Perceptron
Nagarajan
 
Bayesian inversion of deterministic dynamic causal models
Bayesian inversion of deterministic dynamic causal modelsBayesian inversion of deterministic dynamic causal models
Bayesian inversion of deterministic dynamic causal models
khbrodersen
 
Theories of continuous optimization
Theories of continuous optimizationTheories of continuous optimization
Theories of continuous optimization
Olivier Teytaud
 
JAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rules
JAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rulesJAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rules
JAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rules
hirokazutanaka
 
Perceptron 2015.ppt
Perceptron 2015.pptPerceptron 2015.ppt
Perceptron 2015.ppt
SadafAyesha9
 
Artificial Neural Network.pptx
Artificial Neural Network.pptxArtificial Neural Network.pptx
Artificial Neural Network.pptx
shashankbhadouria4
 
Exponent 2
Exponent 2Exponent 2
Exponent 2
Dreams4school
 
Rules of exponents 1
Rules of exponents 1Rules of exponents 1
Rules of exponents 1
lothomas
 
NN-Ch3.PDF
NN-Ch3.PDFNN-Ch3.PDF
NN-Ch3.PDF
gnans Kgnanshek
 

Similar to Hebbian Learning (20)

Neural Networks
Neural NetworksNeural Networks
Neural Networks
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
 
Unit 2
Unit 2Unit 2
Unit 2
 
constructing_generic_algorithms__ben_deane__cppcon_2020.pdf
constructing_generic_algorithms__ben_deane__cppcon_2020.pdfconstructing_generic_algorithms__ben_deane__cppcon_2020.pdf
constructing_generic_algorithms__ben_deane__cppcon_2020.pdf
 
Associative_Memory_Neural_Networks_pptx.pptx
Associative_Memory_Neural_Networks_pptx.pptxAssociative_Memory_Neural_Networks_pptx.pptx
Associative_Memory_Neural_Networks_pptx.pptx
 
Lec10
Lec10Lec10
Lec10
 
CS532L4_Backpropagation.pptx
CS532L4_Backpropagation.pptxCS532L4_Backpropagation.pptx
CS532L4_Backpropagation.pptx
 
NN-Ch3 (1).ppt
NN-Ch3 (1).pptNN-Ch3 (1).ppt
NN-Ch3 (1).ppt
 
CS767_Lecture_04.pptx
CS767_Lecture_04.pptxCS767_Lecture_04.pptx
CS767_Lecture_04.pptx
 
Unit iii update
Unit iii updateUnit iii update
Unit iii update
 
S1 Dualsimplex
S1 DualsimplexS1 Dualsimplex
S1 Dualsimplex
 
Perceptron
PerceptronPerceptron
Perceptron
 
Bayesian inversion of deterministic dynamic causal models
Bayesian inversion of deterministic dynamic causal modelsBayesian inversion of deterministic dynamic causal models
Bayesian inversion of deterministic dynamic causal models
 
Theories of continuous optimization
Theories of continuous optimizationTheories of continuous optimization
Theories of continuous optimization
 
JAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rules
JAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rulesJAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rules
JAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rules
 
Perceptron 2015.ppt
Perceptron 2015.pptPerceptron 2015.ppt
Perceptron 2015.ppt
 
Artificial Neural Network.pptx
Artificial Neural Network.pptxArtificial Neural Network.pptx
Artificial Neural Network.pptx
 
Exponent 2
Exponent 2Exponent 2
Exponent 2
 
Rules of exponents 1
Rules of exponents 1Rules of exponents 1
Rules of exponents 1
 
NN-Ch3.PDF
NN-Ch3.PDFNN-Ch3.PDF
NN-Ch3.PDF
 

More from ESCOM

redes neuronales tipo Som
redes neuronales tipo Somredes neuronales tipo Som
redes neuronales tipo Som
ESCOM
 
redes neuronales Som
redes neuronales Somredes neuronales Som
redes neuronales Som
ESCOM
 
redes neuronales Som Slides
redes neuronales Som Slidesredes neuronales Som Slides
redes neuronales Som Slides
ESCOM
 
red neuronal Som Net
red neuronal Som Netred neuronal Som Net
red neuronal Som Net
ESCOM
 
Self Organinising neural networks
Self Organinising  neural networksSelf Organinising  neural networks
Self Organinising neural networks
ESCOM
 
redes neuronales Kohonen
redes neuronales Kohonenredes neuronales Kohonen
redes neuronales Kohonen
ESCOM
 
Teoria Resonancia Adaptativa
Teoria Resonancia AdaptativaTeoria Resonancia Adaptativa
Teoria Resonancia Adaptativa
ESCOM
 
ejemplo red neuronal Art1
ejemplo red neuronal Art1ejemplo red neuronal Art1
ejemplo red neuronal Art1
ESCOM
 
redes neuronales tipo Art3
redes neuronales tipo Art3redes neuronales tipo Art3
redes neuronales tipo Art3
ESCOM
 
Art2
Art2Art2
Art2
ESCOM
 
Redes neuronales tipo Art
Redes neuronales tipo ArtRedes neuronales tipo Art
Redes neuronales tipo Art
ESCOM
 
Neocognitron
NeocognitronNeocognitron
Neocognitron
ESCOM
 
Neocognitron
NeocognitronNeocognitron
Neocognitron
ESCOM
 
Neocognitron
NeocognitronNeocognitron
Neocognitron
ESCOM
 
Fukushima Cognitron
Fukushima CognitronFukushima Cognitron
Fukushima Cognitron
ESCOM
 
Counterpropagation NETWORK
Counterpropagation NETWORKCounterpropagation NETWORK
Counterpropagation NETWORK
ESCOM
 
Counterpropagation NETWORK
Counterpropagation NETWORKCounterpropagation NETWORK
Counterpropagation NETWORK
ESCOM
 
Counterpropagation
CounterpropagationCounterpropagation
Counterpropagation
ESCOM
 
Teoría de Resonancia Adaptativa Art2 ARTMAP
Teoría de Resonancia Adaptativa Art2 ARTMAPTeoría de Resonancia Adaptativa Art2 ARTMAP
Teoría de Resonancia Adaptativa Art2 ARTMAP
ESCOM
 
Teoría de Resonancia Adaptativa ART1
Teoría de Resonancia Adaptativa ART1Teoría de Resonancia Adaptativa ART1
Teoría de Resonancia Adaptativa ART1
ESCOM
 

More from ESCOM (20)

redes neuronales tipo Som
redes neuronales tipo Somredes neuronales tipo Som
redes neuronales tipo Som
 
redes neuronales Som
redes neuronales Somredes neuronales Som
redes neuronales Som
 
redes neuronales Som Slides
redes neuronales Som Slidesredes neuronales Som Slides
redes neuronales Som Slides
 
red neuronal Som Net
red neuronal Som Netred neuronal Som Net
red neuronal Som Net
 
Self Organinising neural networks
Self Organinising  neural networksSelf Organinising  neural networks
Self Organinising neural networks
 
redes neuronales Kohonen
redes neuronales Kohonenredes neuronales Kohonen
redes neuronales Kohonen
 
Teoria Resonancia Adaptativa
Teoria Resonancia AdaptativaTeoria Resonancia Adaptativa
Teoria Resonancia Adaptativa
 
ejemplo red neuronal Art1
ejemplo red neuronal Art1ejemplo red neuronal Art1
ejemplo red neuronal Art1
 
redes neuronales tipo Art3
redes neuronales tipo Art3redes neuronales tipo Art3
redes neuronales tipo Art3
 
Art2
Art2Art2
Art2
 
Redes neuronales tipo Art
Redes neuronales tipo ArtRedes neuronales tipo Art
Redes neuronales tipo Art
 
Neocognitron
NeocognitronNeocognitron
Neocognitron
 
Neocognitron
NeocognitronNeocognitron
Neocognitron
 
Neocognitron
NeocognitronNeocognitron
Neocognitron
 
Fukushima Cognitron
Fukushima CognitronFukushima Cognitron
Fukushima Cognitron
 
Counterpropagation NETWORK
Counterpropagation NETWORKCounterpropagation NETWORK
Counterpropagation NETWORK
 
Counterpropagation NETWORK
Counterpropagation NETWORKCounterpropagation NETWORK
Counterpropagation NETWORK
 
Counterpropagation
CounterpropagationCounterpropagation
Counterpropagation
 
Teoría de Resonancia Adaptativa Art2 ARTMAP
Teoría de Resonancia Adaptativa Art2 ARTMAPTeoría de Resonancia Adaptativa Art2 ARTMAP
Teoría de Resonancia Adaptativa Art2 ARTMAP
 
Teoría de Resonancia Adaptativa ART1
Teoría de Resonancia Adaptativa ART1Teoría de Resonancia Adaptativa ART1
Teoría de Resonancia Adaptativa ART1
 

Recently uploaded

Get Success with the Latest UiPath UIPATH-ADPV1 Exam Dumps (V11.02) 2024
Get Success with the Latest UiPath UIPATH-ADPV1 Exam Dumps (V11.02) 2024Get Success with the Latest UiPath UIPATH-ADPV1 Exam Dumps (V11.02) 2024
Get Success with the Latest UiPath UIPATH-ADPV1 Exam Dumps (V11.02) 2024
yarusun
 
Keynote given on June 24 for MASSP at Grand Traverse City
Keynote given on June 24 for MASSP at Grand Traverse CityKeynote given on June 24 for MASSP at Grand Traverse City
Keynote given on June 24 for MASSP at Grand Traverse City
PJ Caposey
 
How to Create User Notification in Odoo 17
How to Create User Notification in Odoo 17How to Create User Notification in Odoo 17
How to Create User Notification in Odoo 17
Celine George
 
Erasmus + DISSEMINATION ACTIVITIES Croatia
Erasmus + DISSEMINATION ACTIVITIES CroatiaErasmus + DISSEMINATION ACTIVITIES Croatia
Erasmus + DISSEMINATION ACTIVITIES Croatia
whatchangedhowreflec
 
pol sci Election and Representation Class 11 Notes.pdf
pol sci Election and Representation Class 11 Notes.pdfpol sci Election and Representation Class 11 Notes.pdf
pol sci Election and Representation Class 11 Notes.pdf
BiplabHalder13
 
Creation or Update of a Mandatory Field is Not Set in Odoo 17
Creation or Update of a Mandatory Field is Not Set in Odoo 17Creation or Update of a Mandatory Field is Not Set in Odoo 17
Creation or Update of a Mandatory Field is Not Set in Odoo 17
Celine George
 
Ethiopia and Eritrea Eritrea's journey has been marked by resilience and dete...
Ethiopia and Eritrea Eritrea's journey has been marked by resilience and dete...Ethiopia and Eritrea Eritrea's journey has been marked by resilience and dete...
Ethiopia and Eritrea Eritrea's journey has been marked by resilience and dete...
biruktesfaye27
 
A Quiz on Drug Abuse Awareness by Quizzito
A Quiz on Drug Abuse Awareness by QuizzitoA Quiz on Drug Abuse Awareness by Quizzito
A Quiz on Drug Abuse Awareness by Quizzito
Quizzito The Quiz Society of Gargi College
 
The basics of sentences session 8pptx.pptx
The basics of sentences session 8pptx.pptxThe basics of sentences session 8pptx.pptx
The basics of sentences session 8pptx.pptx
heathfieldcps1
 
Post init hook in the odoo 17 ERP Module
Post init hook in the  odoo 17 ERP ModulePost init hook in the  odoo 17 ERP Module
Post init hook in the odoo 17 ERP Module
Celine George
 
Information and Communication Technology in Education
Information and Communication Technology in EducationInformation and Communication Technology in Education
Information and Communication Technology in Education
MJDuyan
 
IoT (Internet of Things) introduction Notes.pdf
IoT (Internet of Things) introduction Notes.pdfIoT (Internet of Things) introduction Notes.pdf
IoT (Internet of Things) introduction Notes.pdf
roshanranjit222
 
The Science of Learning: implications for modern teaching
The Science of Learning: implications for modern teachingThe Science of Learning: implications for modern teaching
The Science of Learning: implications for modern teaching
Derek Wenmoth
 
Slides Peluncuran Amalan Pemakanan Sihat.pptx
Slides Peluncuran Amalan Pemakanan Sihat.pptxSlides Peluncuran Amalan Pemakanan Sihat.pptx
Slides Peluncuran Amalan Pemakanan Sihat.pptx
shabeluno
 
Diversity Quiz Finals by Quiz Club, IIT Kanpur
Diversity Quiz Finals by Quiz Club, IIT KanpurDiversity Quiz Finals by Quiz Club, IIT Kanpur
Diversity Quiz Finals by Quiz Club, IIT Kanpur
Quiz Club IIT Kanpur
 
Creativity for Innovation and Speechmaking
Creativity for Innovation and SpeechmakingCreativity for Innovation and Speechmaking
Creativity for Innovation and Speechmaking
MattVassar1
 
bryophytes.pptx bsc botany honours second semester
bryophytes.pptx bsc botany honours  second semesterbryophytes.pptx bsc botany honours  second semester
bryophytes.pptx bsc botany honours second semester
Sarojini38
 
Decolonizing Universal Design for Learning
Decolonizing Universal Design for LearningDecolonizing Universal Design for Learning
Decolonizing Universal Design for Learning
Frederic Fovet
 
Observational Learning
Observational Learning Observational Learning
Observational Learning
sanamushtaq922
 
220711130097 Tulip Samanta Concept of Information and Communication Technology
220711130097 Tulip Samanta Concept of Information and Communication Technology220711130097 Tulip Samanta Concept of Information and Communication Technology
220711130097 Tulip Samanta Concept of Information and Communication Technology
Kalna College
 

Recently uploaded (20)

Get Success with the Latest UiPath UIPATH-ADPV1 Exam Dumps (V11.02) 2024
Get Success with the Latest UiPath UIPATH-ADPV1 Exam Dumps (V11.02) 2024Get Success with the Latest UiPath UIPATH-ADPV1 Exam Dumps (V11.02) 2024
Get Success with the Latest UiPath UIPATH-ADPV1 Exam Dumps (V11.02) 2024
 
Keynote given on June 24 for MASSP at Grand Traverse City
Keynote given on June 24 for MASSP at Grand Traverse CityKeynote given on June 24 for MASSP at Grand Traverse City
Keynote given on June 24 for MASSP at Grand Traverse City
 
How to Create User Notification in Odoo 17
How to Create User Notification in Odoo 17How to Create User Notification in Odoo 17
How to Create User Notification in Odoo 17
 
Erasmus + DISSEMINATION ACTIVITIES Croatia
Erasmus + DISSEMINATION ACTIVITIES CroatiaErasmus + DISSEMINATION ACTIVITIES Croatia
Erasmus + DISSEMINATION ACTIVITIES Croatia
 
pol sci Election and Representation Class 11 Notes.pdf
pol sci Election and Representation Class 11 Notes.pdfpol sci Election and Representation Class 11 Notes.pdf
pol sci Election and Representation Class 11 Notes.pdf
 
Creation or Update of a Mandatory Field is Not Set in Odoo 17
Creation or Update of a Mandatory Field is Not Set in Odoo 17Creation or Update of a Mandatory Field is Not Set in Odoo 17
Creation or Update of a Mandatory Field is Not Set in Odoo 17
 
Ethiopia and Eritrea Eritrea's journey has been marked by resilience and dete...
Ethiopia and Eritrea Eritrea's journey has been marked by resilience and dete...Ethiopia and Eritrea Eritrea's journey has been marked by resilience and dete...
Ethiopia and Eritrea Eritrea's journey has been marked by resilience and dete...
 
A Quiz on Drug Abuse Awareness by Quizzito
A Quiz on Drug Abuse Awareness by QuizzitoA Quiz on Drug Abuse Awareness by Quizzito
A Quiz on Drug Abuse Awareness by Quizzito
 
The basics of sentences session 8pptx.pptx
The basics of sentences session 8pptx.pptxThe basics of sentences session 8pptx.pptx
The basics of sentences session 8pptx.pptx
 
Post init hook in the odoo 17 ERP Module
Post init hook in the  odoo 17 ERP ModulePost init hook in the  odoo 17 ERP Module
Post init hook in the odoo 17 ERP Module
 
Information and Communication Technology in Education
Information and Communication Technology in EducationInformation and Communication Technology in Education
Information and Communication Technology in Education
 
IoT (Internet of Things) introduction Notes.pdf
IoT (Internet of Things) introduction Notes.pdfIoT (Internet of Things) introduction Notes.pdf
IoT (Internet of Things) introduction Notes.pdf
 
The Science of Learning: implications for modern teaching
The Science of Learning: implications for modern teachingThe Science of Learning: implications for modern teaching
The Science of Learning: implications for modern teaching
 
Slides Peluncuran Amalan Pemakanan Sihat.pptx
Slides Peluncuran Amalan Pemakanan Sihat.pptxSlides Peluncuran Amalan Pemakanan Sihat.pptx
Slides Peluncuran Amalan Pemakanan Sihat.pptx
 
Diversity Quiz Finals by Quiz Club, IIT Kanpur
Diversity Quiz Finals by Quiz Club, IIT KanpurDiversity Quiz Finals by Quiz Club, IIT Kanpur
Diversity Quiz Finals by Quiz Club, IIT Kanpur
 
Creativity for Innovation and Speechmaking
Creativity for Innovation and SpeechmakingCreativity for Innovation and Speechmaking
Creativity for Innovation and Speechmaking
 
bryophytes.pptx bsc botany honours second semester
bryophytes.pptx bsc botany honours  second semesterbryophytes.pptx bsc botany honours  second semester
bryophytes.pptx bsc botany honours second semester
 
Decolonizing Universal Design for Learning
Decolonizing Universal Design for LearningDecolonizing Universal Design for Learning
Decolonizing Universal Design for Learning
 
Observational Learning
Observational Learning Observational Learning
Observational Learning
 
220711130097 Tulip Samanta Concept of Information and Communication Technology
220711130097 Tulip Samanta Concept of Information and Communication Technology220711130097 Tulip Samanta Concept of Information and Communication Technology
220711130097 Tulip Samanta Concept of Information and Communication Technology
 

Hebbian Learning

  • 1.
  • 2. ART1 Demo Increasing vigilance causes the network to be more selective, to introduce a new prototype when the fit is not good. Try different patterns
  • 4. Hebb’s Postulate “ When an axon of cell A is near enough to excite a cell B and repeatedly or persistently takes part in firing it, some growth process or metabolic change takes place in one or both cells such that A’s efficiency, as one of the cells firing B, is increased.” D. O. Hebb, 1949 A B In other words, when a weight contributes to firing a neuron, the weight is increased. (If the neuron doesn’t fire, then it is not).
  • 5. A B
  • 6. A B
  • 7.
  • 8.
  • 9.
  • 10.
  • 11. Unsupervised Hebbian Learning (aka Associative Learning)
  • 13. Banana Associator Unconditioned Stimulus Conditioned Stimulus Didn’t Pavlov anticipate this?
  • 14. Banana Associator Demo can be toggled
  • 15. Unsupervised Hebb Rule Vector Form: Training Sequence: actual response input
  • 16. Learning Banana Smell Initial Weights: Training Sequence: First Iteration (sight fails, smell present):  = 1 unconditioned (shape) conditioned (smell) a 1   h a r d l i m w 0 p 0 1   w 0   p 1   0.5 – +   h a r d l i m 1 0  0 1  0.5 – +   0 (no banana) = = =
  • 17. Example Second Iteration (sight works, smell present): Third Iteration (sight fails, smell present): Banana will now be detected if either sensor works. a 2   h a r d l i m w 0 p 0 2   w 1   p 2   0.5 – +   h a r d l i m 1 1  0 1  0.5 – +   1 (banana) = = =
  • 18.
  • 19. Hebb Rule with Decay This keeps the weight matrix from growing without bound, which can be demonstrated by setting both a i and p j to 1:
  • 21. Example: Banana Associator First Iteration (sight fails, smell present): Second Iteration (sight works, smell present):  = 0.1  = 1 a 1   h a r d l i m w 0 p 0 1   w 0   p 1   0.5 – +   h a r d l i m 1 0  0 1  0.5 – +   0 (no banana) = = = a 2   h a r d l i m w 0 p 0 2   w 1   p 2   0.5 – +   h a r d l i m 1 1  0 1  0.5 – +   1 (banana) = = =
  • 22. Example Third Iteration (sight fails, smell present):
  • 23. General Decay Demo no decay larger decay w i j m a x   - - - =
  • 24. Problem of Hebb with Decay Associations will be lost if stimuli are not occasionally presented. If a i = 0, then If  = 0, this becomes Therefore the weight decays by 10% at each iteration where there is no stimulus.
  • 25.
  • 27. Instar Operation The instar will be active when or For normalized vectors, the largest inner product occurs when the angle between the weight vector and the input vector is zero -- the input vector is equal to the weight vector. The rows of a weight matrix represent patterns to be recognized. w T 1 p w 1 p  cos b –  =
  • 28. Vector Recognition If we set the instar will only be active when   =  0. If we set the instar will be active for a range of angles. As b is increased, the more patterns there will be (over a wider range of  ) which will activate the instar. b w 1 p – = b w 1 p – > w 1
  • 29. Instar Rule Hebb with Decay Modify so that learning and forgetting will only occur when the neuron is active - Instar Rule: or Vector Form: w i j q   w i j q 1 –    a i q   p j q    a i q   w q 1 –   – + = i j
  • 30. Graphical Representation For the case where the instar is active ( a i = 1): or For the case where the instar is inactive ( a i = 0):
  • 31. Instar Demo weight vector input vector  W
  • 32. Outstar ( Recall Network)
  • 33. Outstar Operation Suppose we want the outstar to recall a certain pattern a * whenever the input p = 1 is presented to the network. Let Then, when p = 1 and the pattern is correctly recalled. The columns of a weight matrix represent patterns to be recalled.
  • 34. Outstar Rule For the instar rule we made the weight decay term of the Hebb rule proportional to the output of the network. For the outstar rule we make the weight decay term proportional to the input of the network. If we make the decay rate  equal to the learning rate  , Vector Form:
  • 42. Hebb Rule Presynaptic Signal Postsynaptic Signal Simplified Form: Supervised Form: Matrix Form: actual output input pattern desired output
  • 43. Batch Operation Matrix Form: (Zero Initial Weights)  W t 1 t 2  t Q p 1 T p 2 T p Q T T P T = = T t 1 t 2  t Q = P p 1 p 2  p Q =
  • 44. Performance Analysis Case I, input patterns are orthogonal. Therefore the network output equals the target: Case II, input patterns are normalized, but not orthogonal. Error term 0 q k  =
  • 45. Example Banana Apple Normalized Prototype Patterns Weight Matrix (Hebb Rule): Tests: Banana Apple
  • 46. Pseudoinverse Rule - (1) Performance Index: Matrix Form: Mean-squared error T t 1 t 2  t Q = P p 1 p 2  p Q = || E || 2 e i j 2 j  i  =
  • 47. Pseudoinverse Rule - (2) Minimize: If an inverse exists for P , F ( W ) can be made zero: When an inverse does not exist F ( W ) can be minimized using the pseudoinverse:
  • 48. Relationship to the Hebb Rule Hebb Rule Pseudoinverse Rule If the prototype patterns are orthonormal: W T P T =
  • 51. Tests 50% Occluded 67% Occluded Noisy Patterns (7 pixels)
  • 53. Spectrum of Hebbian Learning Basic Supervised Rule: Supervised with Learning Rate: Smoothing: Delta Rule: Unsupervised: target actual
  翻译: