![Andrej Karpathy](/img/default-banner.jpg)
- 19
- 2 243 444
Andrej Karpathy
Приєднався 14 кві 2014
Stanford Computer Vision Lab
Pong AI with Policy Gradients
Trained for ~8000 episodes, each episode = ~30 games. Updates were done in batches of 10 episodes, so ~800 updates total. Policy network is a 2-layer neural net connected to raw pixels, with 200 hidden units. Trained with RMSProp and learning rate 1e-4. The final agent does not beat the hard-coded AI consistently, but holds its own. Should be trained longer, with ConvNets, and on GPU.
This is ATARI 2600 Pong version, using OpenAI Gym.
This is ATARI 2600 Pong version, using OpenAI Gym.
Переглядів: 168 433
Відео
Introducing arxiv-sanity
Переглядів 80 тис.8 років тому
Arxiv is great, but there are many papers and we don't have good interfaces for interacting with this large body of work. Enter www.arxiv-sanity.com/ !
CS231n Winter 2016: Lecture 15: Invited Talk by Jeff Dean
Переглядів 37 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 15. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n. Our course website is cs231n.stanford.edu/
CS231n Winter 2016: Lecture 14: Videos and Unsupervised Learning
Переглядів 39 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 14. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n. Our course website is cs231n.stanford.edu/
CS231n Winter 2016: Lecture 13: Segmentation, soft attention, spatial transformers
Переглядів 68 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 13. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n. Our course website is cs231n.stanford.edu/
CS231n Winter 2016: Lecture 12: Deep Learning libraries
Переглядів 46 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 12. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n. Our course website is cs231n.stanford.edu/
CS231n Winter 2016: Lecture 11: ConvNets in practice
Переглядів 47 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 11. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n. Our course website is cs231n.stanford.edu/
CS231n Winter 2016: Lecture 10: Recurrent Neural Networks, Image Captioning, LSTM
Переглядів 112 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 10. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n. Our course website is cs231n.stanford.edu/
CS231n Winter 2016: Lecture 9: Visualization, Deep Dream, Neural Style, Adversarial Examples
Переглядів 62 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 9. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n. Our course website is cs231n.stanford.edu/
CS231n Winter 2016: Lecture 8: Localization and Detection
Переглядів 89 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 8. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n.
CS231n Winter 2016: Lecture 7: Convolutional Neural Networks
Переглядів 164 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 7. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n.
CS231n Winter 2016: Lecture 6: Neural Networks Part 3 / Intro to ConvNets
Переглядів 101 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 6. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n.
CS231n Winter 2016: Lecture 5: Neural Networks Part 2
Переглядів 180 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 5. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n.
CS231n Winter 2016: Lecture 4: Backpropagation, Neural Networks 1
Переглядів 292 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 4. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n.
CS231n Winter 2016: Lecture 3: Linear Classification 2, Optimization
Переглядів 139 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 3. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n.
CS231n Winter 2016: Lecture 2: Data-driven approach, kNN, Linear Classification 1
Переглядів 174 тис.8 років тому
CS231n Winter 2016: Lecture 2: Data-driven approach, kNN, Linear Classification 1
CS231n Winter 2016: Lecture1: Introduction and Historical Context
Переглядів 397 тис.8 років тому
CS231n Winter 2016: Lecture1: Introduction and Historical Context
Large-scale Video Classification with Convolutional Neural Networks, CVPR 2014
Переглядів 42 тис.10 років тому
Large-scale Video Classification with Convolutional Neural Networks, CVPR 2014
Very good lecture, Andrej. Also, the student who interrupted Andrej by saying 'that was what I said' didn't have a good tone and annoyed me.
hey internet, i was here
andrej has the most unique way of telling things and generally they are more intuitive, what a maniac!
yeah completely agree he is just excellent!
Just want to say a big Thank you, this course had helped me prepare the foundation for my PhD back in 2017. Literally watched it multiple times to internalise the concepts
and this is the 100th
this is the 99th comment
27:40
6:00
Amazing Course ! Thank you!
Started my journey and watching this course now. Couldn't attend Stanford University but still good to be in the class in 2024 : )
This is helping me quite a lot, thanks!!!
In 2015, the world didn't know about the problem of batch normalization
1:14:27 - The network never fully converges, but at some point you stopped caring. Because it has been 2 weeks and you are just tired. 😅
You are one of the best Andrej. You make learning so fun, with moments like 27:45 😄 Forever grateful.
This clears lot of doubts I had in my head. Thank you Andrej.
Can't thank Andrej, the cs231n team, Stanford enough. Thoroughly enjoy your lectures. Knowledge is one form of addiction and pleasure and thank you so much for providing it freely. I hope you all enjoy giving it as much as we enjoy receiving it.
1:09:10 it worked :)
does he speaks 1.5x by default 😛
At 39:00, what did he mean by jiggling the scores?
one of the best videos on RNN who explains the code perfectly from sratch.
This content here, is GOLD.
28:00
I love you Andrej ❤
This is a beautiful lecture - gave a very fundamental understanding of backward propagation and its concepts - I see backward propagation correlates to demultiplexing and forward prop corresponds to multiplexing where we are multiplexing the input .
haha. at 1.17.00 Justin says "it's cool but not sure why you would want to generate images"... 🙂
Seems to be a mistake in the "Computing Convolutions: Recap" slide - it says "FFT: Big speedups for small kernels" when it should be "big kernels"?
YOLO is king now hahaha.
when this video was recorded I could not even factor a quadrtic equation. Now I can watch this and follow the math w realetive ease. wow
Day -2 . 3 lectures and counting...
10
9
8
It's still a good way to begin the journey of CV in 2023 , thanks for teaching
Es muy interesante
Tarea 8
Can someone explain why the variance is reduced by 1/2 when using ReLU? Take for example a sample of size=100_000 from a normal distribution N(0, a), pass it through the ReLU and then calculate its variance. Would it be a/2? Moreover, on 45:30 why by setting larger weights the distribution of activations changes shape compared to when using Xavier? I am expecting a flatter distribution compared to Xavier, but not that shape with these peaks on the boundaries. Finally, how these distributions of activations are calculated? Passing many samples through the network with fixed weights?
I wanted to code all that up like you did but I am not able to do it 🥲
i am come from his blog, and he say he put a lot of effort for this class.
Thanks internet
Question! (from an absolute beginner): For the slide shown @24:52, (The bottom example on "Backward pass: guided bakpropagation") Why don't all the negative numbers map to 0? My understanding was that it should automatically be done to deconstruct the image properly. Thnx.
Awesome as always! Is notes available to general public? If so, where do I find it?
She said that she is giving @Andrej a challenge and ChatGPT4 did the same thing in the 14 March 2023 release .
This is the course, I finally understood it. Thank you so so much!!!!
37:00 wrong, each of the four images are a different set of hyper parameters for regularization. As for different initialization - you should look at the paper at figure 4 - The 9 images are different initializations.
37:49 they ask professor why do we look for the minimum of the -log(p) instead of just looking directly for the maximum of p. He says that it makes the math more convenient but I thought the reason was that you want to maximize the joint probability p1*p2*p3*…*pN for your N samples. So instead of maximing a multiplication, you maximize the sum of the logs: log(p1)+log(p2)+log(p3)+… +log(pN).
about reconstruction: reconstruction vs running throw the NNW <=> integral solving vs derive in the sense that every node that was activated gives more information then any node that was not. x^2 + x +15 dx => 2x + 1 => x^2 + x + n where n is the missing info resulted from no surviving the the boundary function.
ty very much!!!
These lectures are amazing. I still go back to review these every few years. Thank you sir.
1:10:10 “can we give a computer one picture and outcomes a description like this… you give the computer one picture - it gives you one sentence… we are not here yet”. This is 2016, and Dr. Fei-Fei Li cannot envision even in her wildest dreams that a few years later, we shall not only overcome this problem but solve the opposite - give a computer a sentence, and it shall create a picture. She was too ‘shy’ to even ask a computer to do such a thing. What is this something that we don’t even dare to ask a computer to do today, but in 2028 computers will do? 😊
So true!
Did you hear about the mathematician mountain climber? He got stuck at a local maximum