<<
Neural Networks Learning the network: Backprop
11-785, Spring 2018 Lecture 4
1 Design exercise
• Input: Binary coded number • Output: One-hot vector
• Input units? • Output units? • Architecture? • Activations?
2 Recap: The MLP can represent any function
• The MLP can be constructed to represent anything • But how do we construct it?
3 Recap: How to learn the function
• By minimizing expected error
4 Recap: Sampling the function
di
Xi
• is unknown, so sample it – Basically, get input-output pairs for a number of samples of input • Many samples , where – Good sampling: the samples of will be drawn from • Estimate function from the samples 5 The Empirical risk
di
Xi
• The expected error is the average error over the entire input space