Kaiming’s ML overview at MIT

1 minute read

Kaiming He joint MIT as associate professor in Feb 2024 and deliveried “Deep Learning Bootcamp” as his first public talk as MIT professor. It’s very pleasant to go over his 1hr talk and here are some notes

DL is all about repensentation.
Alt text LeNet was firstly introduced in 1989 Alt text Conv is local connections with weight sharing. Alt text Pooling is about local invariance Alt text AlexNet in 2012 is big break through after 20yrs. Alt text The net goes wider for more features Alt text and used ReLU for better gradient prop. And used “dropout” Alt text Finally I understand how the visualization works– by back prop a one-hot feature map to pixels!!! Alt text Single most important finding, wow, didn’t think in this way before. Alt text VGG in 2014, too deep to train Alt text Let’s see why from initilization Alt text Vanishing gradients in backprop.
Alt text
Two simple approaches to help w vanishing gradients. Alt text Alt text Inception module was the first time I read paper about NN structures. and varies of inceptions were invented Alt text Alt text Normalization in 2015.
Alt text Batch Norm Alt text Norm in different dimentions Alt text Resnet in 2015 Alt text Can train much deeper nets by fitting the residual Alt text All the key components of DL Alt text RNN is weight sharing in time Alt text It can do deep too
Alt text
and Google Translate in 2016 is an example. Alt text Use CNN for sequence model, which is casual Alt text WaveNet in 2016, used dilation, which is skip connections. Alt text Attention is combining both RNN and CNN for sequency models Alt text Transformers in 2017 Alt text Applied in NLP, we have GPT Alt text AlphaFold is applied to proteins Alt text Applied to general CV/audio Alt text Takeaways, representation and deep/how to go deep Alt text

Tags:

Categories:

Updated: