Neurons and Neural Networks II. Hopfield Network

Neurons and Neural Networks II. Hopfield Network

Neurons and neural networks II. Hopfield network 1 Perceptron recap • key ingredient: adaptivity of the system • unsupervised vs supervised learning • architecture for discrimination: single neuron — perceptron • error function & learning rule • gradient descent learning & divergence • regularisation • learning as inference 2 Interpreting learning as inference So far: optimization wrt. an objective function where 3 Interpreting learning as inference So far: optimization wrt. an objective function where What’s this quirky regularizer, anyway? 3 Interpreting learning as inference Let’s interpret y(x,w) as a probability: 4 Interpreting learning as inference Let’s interpret y(x,w) as a probability: in a compact form: 4 Interpreting learning as inference Let’s interpret y(x,w) as a probability: in a compact form: the likelihood of the input data can be expressed with the original error function function 4 Interpreting learning as inference Let’s interpret y(x,w) as a probability: in a compact form: the likelihood of the input data can be expressed with the original error function function the regularizer has the form of a prior! 4 Interpreting learning as inference Let’s interpret y(x,w) as a probability: in a compact form: the likelihood of the input data can be expressed with the original error function function the regularizer has the form of a prior! what we get in the objective function M(w): the posterior distribution of w: 4 Interpreting learning as inference Relationship between M(w) and the posterior interpretation: minimizing M(w) leads to finding the maximum a posteriori estimate wMP The log probability interpretation of the objective function retains: additivity of errors, while keeping the multiplicativity of probabilities 5 Interpreting learning as inference Properties of the Bayesian estimate 6 Interpreting learning as inference Properties of the Bayesian estimate • The probabilistic interpretation makes our assumptions explicit: by the regularizer we imposed a soft constraint on the learned parameters, which expresses our prior expecations. • An additional plus: beyond getting wMP we get a measure for learned parameter uncertainty 6 Interpreting learning as inference Demo 7 Interpreting learning as inference Demo 7 Interpreting learning as inference Demo 7 Interpreting learning as inference Demo 7 Interpreting learning as inference Making predictions Up to this point the goal was optimization: 8 Interpreting learning as inference Making predictions Up to this point the goal was optimization: 8 Interpreting learning as inference Making predictions Up to this point the goal was optimization: Are we equally confident in the two predictions? 8 Interpreting learning as inference Making predictions Up to this point the goal was optimization: Are we equally confident in the two predictions? The Bayesian answer exploits the probabilistic interpretation: 8 Interpreting learning as inference Calculating Bayesian predictions Predictive probability: 9 Interpreting learning as inference Calculating Bayesian predictions Predictive probability: Likelihood: Weight posterior Partition function: 9 Interpreting learning as inference Calculating Bayesian predictions Predictive probability: Likelihood: Weight posterior Partition function: Finally: 9 Interpreting learning as inference Calculating Bayesian predictions How to solve the integral? 10 Interpreting learning as inference Calculating Bayesian predictions How to solve the integral? Bad news: Monte Carlo integration is needed 10 11 12 Interpreting learning as inference Calculating Bayesian predictions 13 Interpreting learning as inference Calculating Bayesian predictions Original estimate 13 Interpreting learning as inference Calculating Bayesian predictions Original estimate Bayesian estimate 13 Interpreting learning as inference Gaussian approximation Taylor expansion around the MAP estimate 14 Interpreting learning as inference Gaussian approximation Taylor expansion around the MAP estimate The Gaussian approximation: 14 Interpreting learning as inference Gaussian approximation Taylor expansion around the MAP estimate The Gaussian approximation: 14 Neural networks Unsupervised learning Capacity of a single neuron is limited: certain data can only be learned So far, we used a supervised learning paradigm: a teacher was necessary to teach an input-output relation Hopfield networks try to cure both Unsupervised learning: what is it about? Hebb rule: an enlightening example assuming 2 neurons and a weight modification process: This simple rule realizes an associative memory! 15 Neural networks The Hopfield network Architecture: a set of I neurons connected by symmetric synapses of weight wij no self connections: wii=0 output of neuron i: xi Activity rule: Synchronous/ asynchronous update Learning rule: ; 16 Neural networks The Hopfield network Architecture: a set of I neurons connected by symmetric synapses of weight wij no self connections: wii=0 output of neuron i: xi Activity rule: Synchronous/ asynchronous update Learning rule: alternatively, a continuous network can be defined as: ; 16 Neural networks Stability of Hopfield network Are the memories stable? Necessary conditions: symmetric weights; asynchronous update 17 Neural networks Stability of Hopfield network Are the memories stable? Necessary conditions: symmetric weights; asynchronous update 17 Neural networks Stability of Hopfield network Are the memories stable? Necessary conditions: symmetric weights; asynchronous update Robust against perturbation of a subset of weights 17 Neural networks Capacity of Hopfield network How many traces can be memorized by a network of I neurons? 18 Neural networks Capacity of Hopfield network 19.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    41 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us