Unit 1
Unit 1
Unit 1
Unit 1 3 Unit 1 4
Unit 1 6
1
Visual Recognition of People in Images
Machine Learning Approach
General poses, high-
dimensional (30-100 dof)
• It is difficult to explicitly design programs that can recognize
Self-occlusions
people or digits in images
Difficult to segment – Modeling the object structure, the physics, the variability, and
the individual limbs Different
body the image formation process can be very difficult
Loss of Information in the sizes
perspective projection • Instead of designing the programs by hand, we collect many
Partial views
examples that specify the correct outputs for different inputs
Unit 1 9 Unit 1 10
2
Online search engines Displaying the structure of a set of documents using a deep NN
Query:
STREET
Unit 1 13 Unit 1 14
Cities question: Its largest airport was named for a Many digital cameras now detect faces
World War II hero; its second largest, for a World War II
battle
Champions answer 2/3 questions with 85‐95%
accuracy Unit 1 15
3
3D Human Pose – Microsoft’s Kinect (2011) Is Machine Learning Solved?
Lots of success already but…
• Many existing systems lag behind human performance
– Comparatively, see how fast children learn
Unit 1 19 Unit 1 20
Unit 1 21 Unit 1 22
Unit 1 23 Unit 1 24
4
Training and Testing Expectations Training and Testing Process
• Training data: examples we are provided with • In training, based only on the training data, construct a machine that generates
• Testing data: data we will see in the future outputs given inputs
• Training error: average value of loss function on training data – One option is to build machines with small training loss
Test error: the average value of loss function on test data – Ideally we wish the machine to model the main regularities in the data and ignore
the noise. However, if the machine has as many degrees of freedom as the data, it
can fit perfectly. We saw the spline case study
• Our goal is, primarily, not to do well on training data We
already have the answers (outputs) for that data – Avoiding this usually requires model complexity control (regularization)
• We want to perform well on future unseen data • In testing, a new sample is drawn i.i.d. from the same distribution as the
We wish to minimize the test error training data
– This assumption makes it unlikely that important regularities in the test data were
• How can we guarantee it, if we do not have test data? missed in the training data
We will rely on probabilistic assumptions on data variability – We run the machine on new sample and evaluate loss: this is the test error
Unit 1 25 Unit 1 26
Unit 1 27 Unit 1 28
5
Probability Theory Probability Theory
Unit 1 31 Unit 1 32
Sum Rule
Product
Rule
Unit 1 33
Unit 1 35 Unit 1 36
6
Linear Basis Function Models (1) Linear Basis Function Models (2)
Unit 1 37 Unit 1 38
Linear Basis Function Models (3) Linear Basis Function Models (4)
Unit 1 39 Unit 1 40
Linear Basis Function Models (5) Other Basis Function Models (6)
• In a Fourier representation, each basis
function represents a given frequency
and has infinite spatial extent
• Wavelets are localized in both space and
frequency, and by definition are linearly
orthogonal
Unit 1 41 Unit 1 42
7
Maximum Likelihood and Least Maximum Likelihood and Least
Squares (1) Squares (2)
Unit 1 43 Unit 1 44
Unit 1 45 Unit 1 46
Unit 1 47 Unit 1 48
8
Regularized Least Squares (1) Regularized Least Squares (2)
Unit 1 49 Unit 1 50
Unit 1 51 Unit 1 52
9
Multiple Outputs (1) Multiple Outputs (2)
Unit 1 55 Unit 1 56
Thank You
Unit 1 57
10