Train Deep Q-Learning on Atari in PyTorch - Reinforcement Learning DQN Code Tutorial Series p.2

Поділитися
Вставка
  • Опубліковано 16 жов 2024

КОМЕНТАРІ • 20

  • @moikli
    @moikli 3 роки тому +1

    Thanks for making this! I knew nothing about tensors and stuff but watching your videos on repeat has really helped me understand.

    • @brthor1117
      @brthor1117  3 роки тому

      Glad you found it useful! Can you point to a time in the video you are referring to with the epsilon?

    • @moikli
      @moikli 3 роки тому

      @@brthor1117 I quickly edited my comment after posting to remove my question. It's weird that you can still see it. At 16:26 you were working on the line. I ended up updating my epsilon in my act method because it seemed more logical that way. You probably just left it in the main training loop because it was still there from the previous video.

    • @brthor1117
      @brthor1117  3 роки тому

      @@moikli I can only see the edit now 🤷‍♂️. For the epsilon, if you don't multiply the step number by the number of parallel environments, it will decay more slowly with respect to the number of actual environment steps we are taking. It should still work but your agent will improve more slowly. Since I introduced parallel environments in this video it was necessary to make the change.

  • @adelaidaguardadogudino8182
    @adelaidaguardadogudino8182 3 роки тому +4

    thank you very much for your video, do you have the final code to download?

  • @홍성의-i2y
    @홍성의-i2y Рік тому +1

    There is something I do not understand.
    The Nature paper says that it stacks the most recent 4 frames.
    But what you are doing is to play 4 different games in parallel.
    So it is not quite the same with the Nature paper, is it?

  • @ZijianHu-es3ys
    @ZijianHu-es3ys 4 місяці тому

    Thanks for making the video. How can i display the GUI window when i run the observe.py code. (I have finished the trainning)

  • @jacobrohan3257
    @jacobrohan3257 3 роки тому

    I found this very helpful, thanks for making this!

  • @Anonymous-uu8fw
    @Anonymous-uu8fw 8 місяців тому

    15:15 bro why did you bother updating the if statement earlier if you just going to delete it, had me so confused 😭

  • @profit_rules2595
    @profit_rules2595 Рік тому

    I dont get the video with agent playing the game when I run the observe file! can you help me with that @brthor

  • @gustavojhonson7876
    @gustavojhonson7876 3 роки тому

    amazing! thanks from brazil!

    • @brthor1117
      @brthor1117  2 роки тому

      Very cool! You're welcome from USA.

  • @leothebear
    @leothebear 2 роки тому

    hello sir, i am currently trying to learn to train dqn learning on atari using ubuntu but i keep getting the error "BrokenPipeError: [errno 32] Broken pipe". what would you suggest i do to fix this?

  • @SuperRossi86
    @SuperRossi86 2 роки тому

    From where I can get the code?

    • @brthor1117
      @brthor1117  2 роки тому

      check the next video in the series

  • @CHINNOJISANTOSHKUMARNITAP
    @CHINNOJISANTOSHKUMARNITAP Рік тому

    A video on rainbow ensemble please

  • @홍성의-i2y
    @홍성의-i2y Рік тому

    gpu part starts from 36:43

  • @shadmanwadith6848
    @shadmanwadith6848 3 роки тому

    Windows user might get an error :
    FileNotFoundError: Could not find module 'C:\Users\"Your username"\PycharmProjects\"Your Project"\venv\lib\site-packages\atari_py\ale_interface\ale_c.dll' (or one of its dependencies).
    Just follow the steps here,
    stackoverflow.com/questions/63080326/could-not-find-module-atari-py-ale-interface-ale-c-dll-or-one-of-its-dependenc
    put the ale_c.dll in your project file will solve the erro

    • @brthor1117
      @brthor1117  3 роки тому +2

      I highly recommend linux for ai development in general!