DL Lecture8 Autoencoder
DL Lecture8 Autoencoder
DL Lecture8 Autoencoder
Input 𝑥 𝑟 Reconstruction
Autoencoder
𝑥 𝑟
ℎ = 𝑓 𝑥 , 𝑟 = 𝑔 ℎ = 𝑔(𝑓 𝑥 )
Why want to copy input to output
• Not really care about copying
• Historical note: goes back to (LeCun, 1987; Bourlard and Kamp, 1988;
Hinton and Zemel, 1994).
Undercomplete autoencoder
• Constrain the code to have smaller dimension than the input
• Training: minimize a loss function
𝐿 𝑥, 𝑟 = 𝐿(𝑥, 𝑔 𝑓 𝑥 )
𝑥 ℎ 𝑟
Undercomplete autoencoder
• Constrain the code to have smaller dimension than the input
• Training: minimize a loss function
𝐿 𝑥, 𝑟 = 𝐿(𝑥, 𝑔 𝑓 𝑥 )
𝑥 ℎ 𝑟
Probabilistic view of regularizing ℎ
• Suppose we have a probabilistic model 𝑝(ℎ, 𝑥)
• MLE on 𝑥
log 𝑝(𝑥) = log 𝑝(ℎ′ , 𝑥)
ℎ′
Loss Regularization
Sparse autoencoder
• Constrain the code to have sparsity
𝜆 𝜆
• Laplacian prior: 𝑝 ℎ = exp(− ℎ 1)
2 2