Andrej Karpathy
Andrej Karpathy
  • 19
  • 2 243 444
Pong AI with Policy Gradients
Trained for ~8000 episodes, each episode = ~30 games. Updates were done in batches of 10 episodes, so ~800 updates total. Policy network is a 2-layer neural net connected to raw pixels, with 200 hidden units. Trained with RMSProp and learning rate 1e-4. The final agent does not beat the hard-coded AI consistently, but holds its own. Should be trained longer, with ConvNets, and on GPU.
This is ATARI 2600 Pong version, using OpenAI Gym.
Переглядів: 168 433

Відео

Introducing arxiv-sanity
Переглядів 80 тис.8 років тому
Arxiv is great, but there are many papers and we don't have good interfaces for interacting with this large body of work. Enter www.arxiv-sanity.com/ !
CS231n Winter 2016: Lecture 15: Invited Talk by Jeff Dean
Переглядів 37 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 15. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n. Our course website is cs231n.stanford.edu/
CS231n Winter 2016: Lecture 14: Videos and Unsupervised Learning
Переглядів 39 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 14. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n. Our course website is cs231n.stanford.edu/
CS231n Winter 2016: Lecture 13: Segmentation, soft attention, spatial transformers
Переглядів 68 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 13. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n. Our course website is cs231n.stanford.edu/
CS231n Winter 2016: Lecture 12: Deep Learning libraries
Переглядів 46 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 12. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n. Our course website is cs231n.stanford.edu/
CS231n Winter 2016: Lecture 11: ConvNets in practice
Переглядів 47 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 11. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n. Our course website is cs231n.stanford.edu/
CS231n Winter 2016: Lecture 10: Recurrent Neural Networks, Image Captioning, LSTM
Переглядів 112 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 10. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n. Our course website is cs231n.stanford.edu/
CS231n Winter 2016: Lecture 9: Visualization, Deep Dream, Neural Style, Adversarial Examples
Переглядів 62 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 9. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n. Our course website is cs231n.stanford.edu/
CS231n Winter 2016: Lecture 8: Localization and Detection
Переглядів 89 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 8. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n.
CS231n Winter 2016: Lecture 7: Convolutional Neural Networks
Переглядів 164 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 7. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n.
CS231n Winter 2016: Lecture 6: Neural Networks Part 3 / Intro to ConvNets
Переглядів 101 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 6. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n.
CS231n Winter 2016: Lecture 5: Neural Networks Part 2
Переглядів 180 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 5. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n.
CS231n Winter 2016: Lecture 4: Backpropagation, Neural Networks 1
Переглядів 292 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 4. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n.
CS231n Winter 2016: Lecture 3: Linear Classification 2, Optimization
Переглядів 139 тис.8 років тому
Stanford Winter Quarter 2016 class: CS231n: Convolutional Neural Networks for Visual Recognition. Lecture 3. Get in touch on Twitter @cs231n, or on Reddit /r/cs231n.
CS231n Winter 2016: Lecture 2: Data-driven approach, kNN, Linear Classification 1
Переглядів 174 тис.8 років тому
CS231n Winter 2016: Lecture 2: Data-driven approach, kNN, Linear Classification 1
CS231n Winter 2016: Lecture1: Introduction and Historical Context
Переглядів 397 тис.8 років тому
CS231n Winter 2016: Lecture1: Introduction and Historical Context
ConvNet forward pass demo
Переглядів 8 тис.9 років тому
ConvNet forward pass demo
Large-scale Video Classification with Convolutional Neural Networks, CVPR 2014
Переглядів 42 тис.10 років тому
Large-scale Video Classification with Convolutional Neural Networks, CVPR 2014

КОМЕНТАРІ

  • @enestemel9490
    @enestemel9490 28 днів тому

    Very good lecture, Andrej. Also, the student who interrupted Andrej by saying 'that was what I said' didn't have a good tone and annoyed me.

  • @SiD-hq2fo
    @SiD-hq2fo Місяць тому

    hey internet, i was here

  • @susdoge3767
    @susdoge3767 Місяць тому

    andrej has the most unique way of telling things and generally they are more intuitive, what a maniac!

    • @anon-yn9rc
      @anon-yn9rc Місяць тому

      yeah completely agree he is just excellent!

  • @suvarnakadam6557
    @suvarnakadam6557 Місяць тому

    Just want to say a big Thank you, this course had helped me prepare the foundation for my PhD back in 2017. Literally watched it multiple times to internalise the concepts

  • @ayushchaudhary663
    @ayushchaudhary663 Місяць тому

    and this is the 100th

  • @ayushchaudhary663
    @ayushchaudhary663 Місяць тому

    this is the 99th comment

  • @phangb580
    @phangb580 2 місяці тому

    27:40

  • @huongdo1758
    @huongdo1758 2 місяці тому

    6:00

  • @vq8gef32
    @vq8gef32 2 місяці тому

    Amazing Course ! Thank you!

  • @vq8gef32
    @vq8gef32 2 місяці тому

    Started my journey and watching this course now. Couldn't attend Stanford University but still good to be in the class in 2024 : )

  • @twentyeightO1
    @twentyeightO1 2 місяці тому

    This is helping me quite a lot, thanks!!!

  • @akzsh
    @akzsh 3 місяці тому

    In 2015, the world didn't know about the problem of batch normalization

  • @sezaiburakkantarci
    @sezaiburakkantarci 4 місяці тому

    1:14:27 - The network never fully converges, but at some point you stopped caring. Because it has been 2 weeks and you are just tired. 😅

  • @sezaiburakkantarci
    @sezaiburakkantarci 4 місяці тому

    You are one of the best Andrej. You make learning so fun, with moments like 27:45 😄 Forever grateful.

  • @vil9386
    @vil9386 5 місяців тому

    This clears lot of doubts I had in my head. Thank you Andrej.

  • @vil9386
    @vil9386 5 місяців тому

    Can't thank Andrej, the cs231n team, Stanford enough. Thoroughly enjoy your lectures. Knowledge is one form of addiction and pleasure and thank you so much for providing it freely. I hope you all enjoy giving it as much as we enjoy receiving it.

  • @egemeyvecioglu3165
    @egemeyvecioglu3165 5 місяців тому

    1:09:10 it worked :)

  • @vijaypalmanit
    @vijaypalmanit 6 місяців тому

    does he speaks 1.5x by default 😛

  • @Siwon-vv5mi
    @Siwon-vv5mi 7 місяців тому

    At 39:00, what did he mean by jiggling the scores?

  • @mannemsaisivadurgaprasad8987
    @mannemsaisivadurgaprasad8987 7 місяців тому

    one of the best videos on RNN who explains the code perfectly from sratch.

  • @jenishah9825
    @jenishah9825 8 місяців тому

    This content here, is GOLD.

  • @pravachanpatra4012
    @pravachanpatra4012 8 місяців тому

    28:00

  • @lifeisbeautifu1
    @lifeisbeautifu1 10 місяців тому

    I love you Andrej ❤

  • @reachmouli
    @reachmouli 11 місяців тому

    This is a beautiful lecture - gave a very fundamental understanding of backward propagation and its concepts - I see backward propagation correlates to demultiplexing and forward prop corresponds to multiplexing where we are multiplexing the input .

  • @GohOnLeeds
    @GohOnLeeds 11 місяців тому

    haha. at 1.17.00 Justin says "it's cool but not sure why you would want to generate images"... 🙂

  • @GohOnLeeds
    @GohOnLeeds 11 місяців тому

    Seems to be a mistake in the "Computing Convolutions: Recap" slide - it says "FFT: Big speedups for small kernels" when it should be "big kernels"?

  • @zeeshankhanyousafzai5229
    @zeeshankhanyousafzai5229 11 місяців тому

    YOLO is king now hahaha.

  • @padenzimmermann1892
    @padenzimmermann1892 11 місяців тому

    when this video was recorded I could not even factor a quadrtic equation. Now I can watch this and follow the math w realetive ease. wow

  • @piewpok3127
    @piewpok3127 11 місяців тому

    Day -2 . 3 lectures and counting...

  • @user-bp1lc2px6m
    @user-bp1lc2px6m 11 місяців тому

    10

  • @user-bp1lc2px6m
    @user-bp1lc2px6m 11 місяців тому

    9

  • @user-bp1lc2px6m
    @user-bp1lc2px6m 11 місяців тому

    8

  • @user-vs2ej2id3y
    @user-vs2ej2id3y Рік тому

    It's still a good way to begin the journey of CV in 2023 , thanks for teaching

  • @user-xp8xu6he9y
    @user-xp8xu6he9y Рік тому

    Es muy interesante

  • @alexandrogomez5493
    @alexandrogomez5493 Рік тому

    Tarea 8

  • @adosar7261
    @adosar7261 Рік тому

    Can someone explain why the variance is reduced by 1/2 when using ReLU? Take for example a sample of size=100_000 from a normal distribution N(0, a), pass it through the ReLU and then calculate its variance. Would it be a/2? Moreover, on 45:30 why by setting larger weights the distribution of activations changes shape compared to when using Xavier? I am expecting a flatter distribution compared to Xavier, but not that shape with these peaks on the boundaries. Finally, how these distributions of activations are calculated? Passing many samples through the network with fixed weights?

  • @tilakrajchoubey5534
    @tilakrajchoubey5534 Рік тому

    I wanted to code all that up like you did but I am not able to do it 🥲

  • @budiardjo6610
    @budiardjo6610 Рік тому

    i am come from his blog, and he say he put a lot of effort for this class.

  • @kemalware4912
    @kemalware4912 Рік тому

    Thanks internet

  • @heyloo0511
    @heyloo0511 Рік тому

    Question! (from an absolute beginner): For the slide shown @24:52, (The bottom example on "Backward pass: guided bakpropagation") Why don't all the negative numbers map to 0? My understanding was that it should automatically be done to deconstruct the image properly. Thnx.

  • @bhavinmoriya9216
    @bhavinmoriya9216 Рік тому

    Awesome as always! Is notes available to general public? If so, where do I find it?

  • @rahularyansharma
    @rahularyansharma Рік тому

    She said that she is giving @Andrej a challenge and ChatGPT4 did the same thing in the 14 March 2023 release .

  • @sherifbadawy8188
    @sherifbadawy8188 Рік тому

    This is the course, I finally understood it. Thank you so so much!!!!

  • @AvielLivay
    @AvielLivay Рік тому

    37:00 wrong, each of the four images are a different set of hyper parameters for regularization. As for different initialization - you should look at the paper at figure 4 - The 9 images are different initializations.

  • @AvielLivay
    @AvielLivay Рік тому

    37:49 they ask professor why do we look for the minimum of the -log(p) instead of just looking directly for the maximum of p. He says that it makes the math more convenient but I thought the reason was that you want to maximize the joint probability p1*p2*p3*…*pN for your N samples. So instead of maximing a multiplication, you maximize the sum of the logs: log(p1)+log(p2)+log(p3)+… +log(pN).

  • @neriyacohen7805
    @neriyacohen7805 Рік тому

    about reconstruction: reconstruction vs running throw the NNW <=> integral solving vs derive in the sense that every node that was activated gives more information then any node that was not. x^2 + x +15 dx => 2x + 1 => x^2 + x + n where n is the missing info resulted from no surviving the the boundary function.

  • @eduardtsuranov712
    @eduardtsuranov712 Рік тому

    ty very much!!!

  • @NehadHirmiz
    @NehadHirmiz Рік тому

    These lectures are amazing. I still go back to review these every few years. Thank you sir.

  • @AvielLivay
    @AvielLivay Рік тому

    1:10:10 “can we give a computer one picture and outcomes a description like this… you give the computer one picture - it gives you one sentence… we are not here yet”. This is 2016, and Dr. Fei-Fei Li cannot envision even in her wildest dreams that a few years later, we shall not only overcome this problem but solve the opposite - give a computer a sentence, and it shall create a picture. She was too ‘shy’ to even ask a computer to do such a thing. What is this something that we don’t even dare to ask a computer to do today, but in 2028 computers will do? 😊

  • @jonathanr4242
    @jonathanr4242 Рік тому

    Did you hear about the mathematician mountain climber? He got stuck at a local maximum