- 01L – Gradient descent and the backpropagation algorithm

01L – Gradient descent and the backpropagation algorithm

Course website: http://bit.ly/DLSP21-web
Playlist: http://bit.ly/DLSP21-YouTube
Speaker: Yann LeCun

Chapters
00:00:00 – Supervised learning
00:03:43 – Parametrised models
00:07:23 – Block diagram
00:08:55 – Loss function, average loss
00:12:23 – Gradient descent
00:30:47 – Traditional neural net...
Course website: http://bit.ly/DLSP21-web
Playlist: http://bit.ly/DLSP21-YouTube
Speaker: Yann LeCun

Chapters
00:00:00 – Supervised learning
00:03:43 – Parametrised models
00:07:23 – Block diagram
00:08:55 – Loss function, average loss
00:12:23 – Gradient descent
00:30:47 – Traditional neural nets
00:35:07 – Backprop through a non-linear function
00:40:41 – Backprop through a weighted sum
00:50:55 – PyTorch implementation
00:57:18 – Backprop through a functional module
01:05:08 – Backprop through a functional module
01:12:15 – Backprop in practice
01:33:15 – Learning representations
01:42:14 – Shallow networks are universal approximators!
01:47:25 – Multilayer architectures == compositional structure of data

#PyTorch #NYU #Yann LeCun #Deep Learning #neural networks
– Supervised learning - 01L – Gradient descent and the backpropagation algorithm

– Supervised learning

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:00:00 - 00:03:43
@Alfredo Canziani Oh right! sorry. It's at - 01L – Gradient descent and the backpropagation algorithm

@Alfredo Canziani Oh right! sorry. It's at

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:01:18 - 01:51:04
– Parametrised models - 01L – Gradient descent and the backpropagation algorithm

– Parametrised models

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:03:43 - 00:07:23
– Block diagram - 01L – Gradient descent and the backpropagation algorithm

– Block diagram

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:07:23 - 00:08:55
– Loss function, average loss - 01L – Gradient descent and the backpropagation algorithm

– Loss function, average loss

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:08:55 - 00:12:23
Discussion on stochastic gradient descent () and with adams (1:16:15) are great. General misconception. - 01L – Gradient descent and the backpropagation algorithm

Discussion on stochastic gradient descent () and with adams (1:16:15) are great. General misconception.

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:12:23 - 01:51:04
– Gradient descent - 01L – Gradient descent and the backpropagation algorithm

– Gradient descent

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:12:23 - 00:30:47
Doesn't he contradict himself? First he mentions that smaller batches are better  (I assume that by "better" he meant model quality) in most cases, and a few seconds later he says that it's just a hardware matter. - 01L – Gradient descent and the backpropagation algorithm

Doesn't he contradict himself? First he mentions that smaller batches are better (I assume that by "better" he meant model quality) in most cases, and a few seconds later he says that it's just a hardware matter.

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:20:03 - 01:51:04
Hi Alfredo, in  Yann mentioned "objective function need to be Continuous mostly and differentiable almost everywhere". What does he mean? isn't the function differentiable is always continuous? also is there a function where some part only differentiable? Can someone give me one example in deep learning functions? pls help me out. - 01L – Gradient descent and the backpropagation algorithm

Hi Alfredo, in Yann mentioned "objective function need to be Continuous mostly and differentiable almost everywhere". What does he mean? isn't the function differentiable is always continuous? also is there a function where some part only differentiable? Can someone give me one example in deep learning functions? pls help me out.

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:20:58 - 01:51:04
My question was from - 01L – Gradient descent and the backpropagation algorithm

My question was from

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:30:17 - 01:51:04
– Traditional neural nets - 01L – Gradient descent and the backpropagation algorithm

– Traditional neural nets

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:30:47 - 00:35:07
@Alfredo Canziani ah, sorry. Was just to add on, at ~  when Prof. LeCun explains why people don't like to refer to the units as 'neurons' persay - 01L – Gradient descent and the backpropagation algorithm

@Alfredo Canziani ah, sorry. Was just to add on, at ~ when Prof. LeCun explains why people don't like to refer to the units as 'neurons' persay

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:31:25 - 01:51:04
– Backprop through a non-linear function - 01L – Gradient descent and the backpropagation algorithm

– Backprop through a non-linear function

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:35:07 - 00:40:41
At the  section - is the purpose of using back propagation to find the derivative of the cost function wrt z to find the best direction to "move"? I've only gotten through half of the lecture so forgive me if this is answered later - 01L – Gradient descent and the backpropagation algorithm

At the section - is the purpose of using back propagation to find the derivative of the cost function wrt z to find the best direction to "move"? I've only gotten through half of the lecture so forgive me if this is answered later

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:40:41 - 01:51:04
– Backprop through a weighted sum - 01L – Gradient descent and the backpropagation algorithm

– Backprop through a weighted sum

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:40:41 - 00:50:55
– PyTorch implementation - 01L – Gradient descent and the backpropagation algorithm

– PyTorch implementation

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:50:55 - 00:57:18
About the code in PyTorch... ( in the video)... the code instantiates the mynet class and stores the reference in model variable... but nowhere it calls the "forward" method... so how does the out variable receive any output from the model object? Is there some Pytorch magic which is not explained here ? - 01L – Gradient descent and the backpropagation algorithm

About the code in PyTorch... ( in the video)... the code instantiates the mynet class and stores the reference in model variable... but nowhere it calls the "forward" method... so how does the out variable receive any output from the model object? Is there some Pytorch magic which is not explained here ?

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:51:00 - 01:51:04
shouldn't it be self.m0(z0) as it takes in the flattened input? - 01L – Gradient descent and the backpropagation algorithm

shouldn't it be self.m0(z0) as it takes in the flattened input?

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:51:05 - 01:51:04
@Alfredo Canziani Actually I asked that before I watched it at , Regards 🤞 - 01L – Gradient descent and the backpropagation algorithm

@Alfredo Canziani Actually I asked that before I watched it at , Regards 🤞

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
00:54:30 - 01:51:04
Just FYI, at  Yann correctly says dc/dzg, but the diagram has dc/zg.  Should that also be dc/dwg and dc/dwf? - 01L – Gradient descent and the backpropagation algorithm

Just FYI, at Yann correctly says dc/dzg, but the diagram has dc/zg. Should that also be dc/dwg and dc/dwf?

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
01:01:00 - 01:51:04
– Backprop through a functional module - 01L – Gradient descent and the backpropagation algorithm

– Backprop through a functional module

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
01:05:08 - 01:12:15
At  Yann is explaining the two jacobians, but I was having trouble getting the intuition.  Then I realized that the first jacobian was getting the gradient to modify the weights w[k+1] for function z[k+1] and the second jacobian was back propagating the gradient to function z[k] which can then be used to calculate the gradient at k for yet another jacobian to adjust weights w[k].  So one jacobian is for the parameters and the other is for the state since both the parameter variable and state variable are column vectors.  Yann explains it really well.  I'm amazed that I seem to be understanding this complicated mix of symbols and logic.  Thank you. - 01L – Gradient descent and the backpropagation algorithm

At Yann is explaining the two jacobians, but I was having trouble getting the intuition. Then I realized that the first jacobian was getting the gradient to modify the weights w[k+1] for function z[k+1] and the second jacobian was back propagating the gradient to function z[k] which can then be used to calculate the gradient at k for yet another jacobian to adjust weights w[k]. So one jacobian is for the parameters and the other is for the state since both the parameter variable and state variable are column vectors. Yann explains it really well. I'm amazed that I seem to be understanding this complicated mix of symbols and logic. Thank you.

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
01:05:40 - 01:51:04
– Backprop in practice - 01L – Gradient descent and the backpropagation algorithm

– Backprop in practice

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
01:12:15 - 01:33:15
Does the trick explained in normalizing training samples () applies also to convolutional neural networks? - 01L – Gradient descent and the backpropagation algorithm

Does the trick explained in normalizing training samples () applies also to convolutional neural networks?

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
01:20:00 - 01:51:04
It's just after . - 01L – Gradient descent and the backpropagation algorithm

It's just after .

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
01:27:00 - 01:51:04
– Learning representations - 01L – Gradient descent and the backpropagation algorithm

– Learning representations

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
01:33:15 - 01:42:14
– Shallow networks are universal approximators! - 01L – Gradient descent and the backpropagation algorithm

– Shallow networks are universal approximators!

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
01:42:14 - 01:47:25
– Multilayer architectures == compositional structure of data - 01L – Gradient descent and the backpropagation algorithm

– Multilayer architectures == compositional structure of data

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
01:47:25 - 01:51:04
I thought that haar-like features were not that recognizable. () - 01L – Gradient descent and the backpropagation algorithm

I thought that haar-like features were not that recognizable. ()

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
01:48:00 - 01:51:04
.  (around ) - 01L – Gradient descent and the backpropagation algorithm

. (around )

01L – Gradient descent and the backpropagation algorithm
2021年07月14日 
01:50:00 - 01:51:04

Alfredo Canziani

※本サイトに掲載されているチャンネル情報や動画情報はYouTube公式のAPIを使って取得・表示しています。

Timetable

動画タイムテーブル