dtransposed
dtransposed
  • 7
  • 28 744
The Game Of Life in Rust
Short video where I present the implementation of the Game of Life in Rust.
Links:
- The GitHub repository: github.com/dtransposed/code_videos/tree/main/02_The_Game_Of_Life_In_Rust/game_of_life
- My website: dtransposed.github.io
- My Twitter: dtransposed
Sources:
- Rustlings: github.com/rust-lang/rustlings
- Visualisations used: ua-cam.com/video/xP5-iIeKXE8/v-deo.html
- Computerspielemuseum Berlin: www.visitberlin.de/en/computerspielemuseum-computer-games-museum
- Diagram for the rules used in the video: www.researchgate.net/figure/Rules-of-Conways-Game-of-Life_fig5_339605473
Timestamps:
0:00 Intro
1:17 Introducing the Game of Life
8:55 Running few simulations, discussing results
15:08 Commenting on the features of Rust language
21:37 Outro
Переглядів: 547

Відео

Diffusion Models - Live Coding Tutorial
Переглядів 26 тис.Рік тому
This is my live (to the most extent) coding video, where I implement from a scratch a diffusion model that generates 32 x 32 RGB images. The tutorial assumes a basic knowledge of deep learning and Python. Links: - The Jupiter notebook built in this video: github.com/dtransposed/code_videos/blob/main/01_Diffusion_Models_Tutorial - My website: dtransposed.github.io - My Twitter: dtran...
Damian Bogunowicz - Sim2Real Industrial Robots
Переглядів 7534 роки тому
Damian Bogunowicz - Sim2Real Industrial Robots
G.E.A.R - BC
Переглядів 2055 років тому
G.E.A.R - BC
G.E.A.R - Heuristic
Переглядів 1905 років тому
G.E.A.R - Heuristic
G.E.A.R - PPOSegNet
Переглядів 3405 років тому
G.E.A.R - PPOSegNet

КОМЕНТАРІ

  • @maheshreddy0
    @maheshreddy0 3 місяці тому

    paper link?

  • @outroutono4937
    @outroutono4937 7 місяців тому

    I have looked almost every video on this subject and this is by far the best approach, it's simple enough to be well understood but it gives all the tools to built more advanced models. I wish you could do a remake of this one because sometimes the code snippet is out of frame and sometimes its hard to read because of the font size. Thx a lot for this upload!

  • @777chichi
    @777chichi 9 місяців тому

    Thanks a lot! I really appreciate. This tutorial explain clearly. Awesome! Hope to see more tutorial vedios on your youtube channel, thanks.

  • @duyquangnguyen2664
    @duyquangnguyen2664 Рік тому

    Thank for this video. Can you make video about apply high resolution for this project ?

  • @Folotilo15
    @Folotilo15 Рік тому

    Thanks a lot for your tutorial!

  • @playmaker2404
    @playmaker2404 Рік тому

    can u say why output was not as fascinating and what can be done from here to make output clearer @dtransposed79

  • @paneercheeseparatha
    @paneercheeseparatha Рік тому

    You should have zoomed in the screen more so that its visible properly. Still appreciate your efforts! Nice vid.

  • @danielfirebanks4973
    @danielfirebanks4973 Рік тому

    Good tutorial, just wished that we could see the screen while you're coding, as most of the new lines you added were off-screen :/ Keep it up!

  • @WaiPanTam
    @WaiPanTam Рік тому

    Thanks man, I really appreciate your work

  • @kanakraj3198
    @kanakraj3198 Рік тому

    Great tutorial. Thanks for sharing. Please make slightly advanced tutorials, like Conditonal (Image or Text) Generation of Images using Diffusion. I see that there are very few advanced tutorials by any UA-camr.

  • @bbbaaa9421
    @bbbaaa9421 Рік тому

    Thanks for sharing your work with us, Appreciate!,

  • @anshumansinha5874
    @anshumansinha5874 Рік тому

    Hi, thanks for the video. But can you explain the part on how you introduce the positional encoding to the network? Also, can this model work for a feed forward neural network rather than a U-net ?

    • @dtransposed79
      @dtransposed79 Рік тому

      Positional encodings in this paper directly mimic those introduced in the "Attention Is All You Need" paper. There are plenty of resources online that explain how that works. In terms of the architecture, in theory, you could probably use any encoder-decoder architecture I think. But for images, UNet is the most fitting.

  • @nqvst
    @nqvst Рік тому

    Coming from a programming background, I always find it very strange to name variables by generic Greek letters or just X, Y. I am not criticizing your video specifically, it is a pattern that is very wide spread. But for example, you are naming the first parameter to the forward_diffusion function "x0". is it to save space? is it because you think it is easier to reference it from the mathematical formulas? In my mind it would be much more clear if "x0" would be named "image". or am I misunderstanding your explanation maybe. As I mentioned, I don't think your video is bad. I'm just curious as to why it is so common that code related to machine learning is generally so generically named.

    • @dtransposed79
      @dtransposed79 Рік тому

      Interesting comment. I agree - some people indeed use more "mathematical" names, and others use more generic ones. Using the "mathematical" names comes from the fact that many of the ML code you can find online implements a logic showcased in a research paper. Since ML research borrows from the mathematical notation, the it is often convenient for the code to use the same notation, as long as they have the same context (read the paper, understand what the symbols mean). If you are confused, I would advice you to read the paper and even if you are confused by any concept, just try to grasp the high level meaning of the symbols. This would definitely help you with reading (and writing) your ML code in the future!

    • @nqvst
      @nqvst Рік тому

      @@dtransposed79 Thanks for the reply! Yes it makes sens. if you understand the concept from reading the equations, it is more convenient to reuse the notation in the code. And while following along this video i realized that some of the variable names gets really long if they are to be considered "good" variable names. betas -> noise_amount alphas -> preserved_image_data alpha_hat_t -> cumulative_preserved_image_data_at_step i think I'm just frustrated over not being fluent in the math language. anyways, thanks for the video!

  • @thepresistence5935
    @thepresistence5935 Рік тому

    Awesome one!

  • @chiscoduran9517
    @chiscoduran9517 Рік тому

    Can you make a Image to Image tutorial?

    • @dtransposed79
      @dtransposed79 Рік тому

      Could you be more concrete? Image-to-Image can mean multiple things.

    • @chiscoduran9517
      @chiscoduran9517 Рік тому

      @@dtransposed79 for example a model capable of change colors to certain objects in an image, where input is an image and put is the same image with changes.

  • @heera_ai
    @heera_ai Рік тому

    A great tutorial to start with!!!

  • @МихаилЮрков-т1э

    Thanks for tutorial. Why posterior_variance_t = betas_t? Shouldn't it be equal to betas_t*(1 - alphas_cumprod_t_minus_1)/(1 - alphas_cumprod_t) according [Lil' Log]?

    • @dtransposed79
      @dtransposed79 Рік тому

      Excellent question. Please refer to the original paper: arxiv.org/abs/2006.11239 Section 3.2. The short version: those two are the extreme values that we can set the posterior to. The choice will depend on the assumptions on x_0. My choice assumes that x_0 is sampled from Gaussian ~ N(0,1), while the other choice is optimal for x_0 deterministically set to one point.

    • @МихаилЮрков-т1э
      @МихаилЮрков-т1э Рік тому

      @@dtransposed79 Yes, it's clear now. Thanks for the detailed answer.

  • @aleksandrrybnikov8701
    @aleksandrrybnikov8701 Рік тому

    Hi, Damian! Nice videos!

  • @brunokemmer
    @brunokemmer Рік тому

    Is there a difference between `result = alpha_hat.gather(-1, t)` and `result = alpha_hat[t]` ?

    • @dontitube1394
      @dontitube1394 Рік тому

      No there is not, atleast for this kind of case. But for more information you can look at the documantation of torch.gather, which even states the equivalant indexing of arrays.

    • @dtransposed79
      @dtransposed79 Рік тому

      @@dontitube1394 Yeah., that's right. Nevertheless, I suggest learning and using torch.gather, It is a really useful, powerful and efficient function.

  • @adeolaogunleye7965
    @adeolaogunleye7965 Рік тому

    Thanks man, I really appreciate your work

  • @VitaliyHAN
    @VitaliyHAN Рік тому

    Better font, but still can’t read not only the phone, that is main content consuming device, but even on my 13 inch MacBook. God bless I have 55 inch tv I can watch on. Even with such struggles I will continue to watch such a diamond video! Thanks for video! Great content!

    • @dtransposed79
      @dtransposed79 Рік тому

      Thank you for your comment!

    • @dontitube1394
      @dontitube1394 Рік тому

      @@dtransposed79 One more tip, @ 34:36 and sometime so on, i cannot read the code you were writing. I mean litterly it is not in the video, but very informativ video.

    • @dtransposed79
      @dtransposed79 Рік тому

      @@dontitube1394 Yeah. I think I will not be changing it now. A bit of a hiccup, but you can always look the code up in the attached notebook. Sorry for that tho.

    • @dontitube1394
      @dontitube1394 Рік тому

      @@dtransposed79 yeah no worries, it was more ment as a tip for future videos

  • @abdelalizyn5859
    @abdelalizyn5859 3 роки тому

    Great video, well explained!

  • @DroneMesh
    @DroneMesh 5 років тому

    I'm surprised why these videos never get views.. Is it really such a small field?

  • @yakupgorur
    @yakupgorur 5 років тому

    Good job! very nice illustration of usage area of RL