Language Models For Software Developers in 17 Minutes

Поділитися
Вставка
  • Опубліковано 22 кві 2023
  • A fast guide on how you can incorporate state of the art language models into your application without relying on a pay-by-request service like OpenAI.
    Code from the video: github.com/Me163/youtube/tree...
    Hugging Face Course: huggingface.co/learn/nlp-cour...
    Camera: Canon EOS R5 amzn.to/3CCrxzl
    Monitor: Dell U4914DW 49in amzn.to/3MJV1jx
    SSD for Video Editing: VectoTech Rapid 8TB amzn.to/3hXz9TM
    Microphone 1: Rode NT1-A amzn.to/3vWM4gL
    Microphone 2: Seinheiser 416 amzn.to/3Fkti60
    Microphone Interface: Focusrite Clarett+ 2Pre amzn.to/3J5dy7S
    Tripod: JOBY GorillaPod 5K amzn.to/3JaPxMA
    Mouse: Razer DeathAdder amzn.to/3J9fYCf
    Keyboard (sometimes): Keychron Q1 amzn.to/3YkJNrB
    Computer: 2021 Macbook Pro amzn.to/3J7FXtW
    Lens: Canon RF24mm F1.8 Macro is STM Lens amzn.to/3UUs1bB
    Caffeine: High Brew Cold Brew Coffee amzn.to/3hXyx0q
    More Caffeine: Monster Energy Juice, Pipeline Punch amzn.to/3Czmfox
    Building A Second Brain book: amzn.to/3cIShWf
  • Наука та технологія

КОМЕНТАРІ • 147

  • @lebogang_nkoane
    @lebogang_nkoane Рік тому +95

    Possibly the best tutorial on how to plug and play with the models. Much appreciated, I gained valuable insight. Respect!✊

  • @clivebird5729
    @clivebird5729 11 місяців тому +20

    Thank you. This was a very impressive, hype free, presentation. 10 out of 10.

  • @joerua470
    @joerua470 10 місяців тому +5

    Thanks for the best plug and play model video out there. Super short and right to the point.

    • @codetothemoon
      @codetothemoon  10 місяців тому

      thanks for watching, glad you got something out of it!

  • @AdrianMark
    @AdrianMark 11 місяців тому +1

    Thanks so much. I just read through the privateGPT py files, and I don't have any python experience. This video helped me understand the reasoning behind that code.

  • @TheMzbac
    @TheMzbac 11 місяців тому +7

    As always, very informative and clear. Explains the essentials.

  • @siddharthmanumusic
    @siddharthmanumusic 10 місяців тому +1

    Wow man, this is so good and simple! Didn't know it were this easy to get a model and run it!

    • @codetothemoon
      @codetothemoon  10 місяців тому

      nice, really happy you got something out of the video!

  • @Mustafa-099
    @Mustafa-099 10 місяців тому +1

    Wow man you have put so much effort into this tutorial!
    I love it, thank you :)

    • @codetothemoon
      @codetothemoon  10 місяців тому

      i did, thanks for watching, glad you got something out of it!

  • @mu3076
    @mu3076 11 місяців тому +1

    *I just found myself a Goldmine, definitely subscribing*

  • @subem81
    @subem81 11 місяців тому +6

    This was really great! Would love to see something on how to LoRA train a model for text generation. Would be really cool!

    • @codetothemoon
      @codetothemoon  11 місяців тому +3

      thanks glad you got something out of it! fine-tuning in general is definitely something I'd like to cover if I can, as it's likely something most productized uses of LLMs will benefit from...

  • @simonabunker
    @simonabunker 11 місяців тому +23

    This is a really great tutorial! The obvious followup would be fine tuning an LLM with your own data. Or maybe using LangChain to do it?

    • @codetothemoon
      @codetothemoon  11 місяців тому +8

      thanks, and I agree! I'd love to make videos on both of those topics.

    • @FabrizioPalmas
      @FabrizioPalmas 10 місяців тому +1

      @@codetothemoono it then 😁 (great video)

  • @ohorit
    @ohorit 7 місяців тому +1

    Excellent!

  • @powpowpony9920
    @powpowpony9920 10 місяців тому

    Great tutorial!

  • @ibriarmiroi
    @ibriarmiroi 10 місяців тому +1

    Thank you for your content
    I would like to ask you to make a video on the web app and the new developments in that regard.

    • @codetothemoon
      @codetothemoon  10 місяців тому +1

      thanks for watching! I will definitely be making more content around Rust full stack web applications.

  • @chris.dillon
    @chris.dillon 9 місяців тому

    Your REPL is a manual ML evaluation. Many devs don't like manual stuff and so eventually we will all realize that figuring out if it works (like asking it the questions the base model flubbed) is manual testing. It's sort of similar to functional testing. "I just coded up something but how do I know it works?". We have a choice to make on whether to quickly manually test it with the massive trade-off of: you cannot possibly ask it every question. Put another way: ask it 10,000 questions every time you change something (code/scripts/model/parameters). Give it a score. The score is still not enough without context.
    The model itself it not just bigger for the sake of being bigger but would be improved and measured. We would be attracted to better performing models. In other words, we expect to consume software and models that are well-functioning and performant. Those things use automation.
    We've been doing this for a long time with plain old software. In ML, this is the new bit. The data and how the result is not deterministic. This is different than plain software.

  • @chrispycryptic
    @chrispycryptic 11 місяців тому +8

    This is amazing! Thank you so much! It is not some "black box" that so many of these doomsayers bleat incessantly, but rather the closest we can really get to a superpower for those who thoroughly understand them!

    • @codetothemoon
      @codetothemoon  11 місяців тому +1

      You're very welcome, glad you got something out of it! 😎

    • @otakarbeinhauer
      @otakarbeinhauer 11 місяців тому +3

      I'm not a doomsayer, but I agree with the "black box" tag that AI models get. I don't understand how this video proved to you that it is not a "black box". The guy making the video was often literally surprised by the answers he got from the language model. And when you're unable to predict the answer, that's basically a "black box" by definition.
      You probably think that "black box" means we don't understand how the language models work. That is not correct. What is meant by the phrase is, that we don't know what those models will return. I hope i clarified things for you.

  • @Christian-op1ss
    @Christian-op1ss Рік тому +3

    Hi, thanks for the nice video! A question, in your Rust video you refer to, you load libtorch, or pytorch. However if you build that program in Rust and give it to someone, would they not also have to install libtorch on their machine and set their environment? And if so, do you know of a way to embed libtorch or pytorch?

  • @normanlove222
    @normanlove222 11 місяців тому

    I would love if you did a video taking this one step further and shopw how to connect to our local data using llama index

  • @sbx1720
    @sbx1720 Рік тому +2

    Great video, thanks!

  • @juanmadelaflor6406
    @juanmadelaflor6406 Рік тому

    Is this possible on a node environment?Thanks.

  • @FabianBarajas
    @FabianBarajas 11 місяців тому +1

    Dude, this is amazing

    • @codetothemoon
      @codetothemoon  11 місяців тому

      Nice, glad you got something out of it!

  • @bruninhohenrri
    @bruninhohenrri 10 місяців тому

    How can i load an list of "intentions" for a language model to classify ? Like when i send some question like: "Please turn on the lights" i would like that the language model classify the "turn_on" intentions as true with the entity "lights"

  • @DreamsAPI
    @DreamsAPI 10 місяців тому +1

    Subscribed. Thank you

    • @codetothemoon
      @codetothemoon  10 місяців тому

      fantastic, really happy to have you on board!

  • @Metruzanca
    @Metruzanca Рік тому +5

    This is excelent. Just the other day I was wondering if I could take a markdown wiki (or obsidian vault) and feed it to a LM to ask questions about it.
    Specifically I wanted to take a messy wiki from work and giving it to a LM and then asking the LM questions. Then we can improve the wiki and get an ever better chatbot that can answer questions. Then can do something like feed it the python and django docs as a baseline.

    • @codetothemoon
      @codetothemoon  Рік тому +1

      nice, yes I think what you describe is definitely achievable!

  • @staystealth
    @staystealth Рік тому +3

    thank you so much for this

    • @codetothemoon
      @codetothemoon  Рік тому

      thanks for watching, glad you got something out of it!

  • @kellybrower7634
    @kellybrower7634 10 місяців тому +1

    I’m curious about your vim workflow. Nice colorings etc. also how do you use the keyboard to open a penv terminal automatically sized side by side?

    • @codetothemoon
      @codetothemoon  10 місяців тому

      thanks, actually its doom emacs! To open the python terminal you can just use M-x then select "run python". I'm using the doom-monokai-pro theme.

  • @papa-pete
    @papa-pete 10 місяців тому

    Can I use tensorflow instead of pytorch?

  • @dwivedys
    @dwivedys 10 місяців тому +1

    For running transformers on windows machine - do I need pytorch or tensorflow or both installed?

    • @codetothemoon
      @codetothemoon  10 місяців тому

      you'd definitely need one or the other (this particular example uses pytorch but I think you can just replace "pt" with "tf" for tensorflow). Not sure about any of the specifics on getting it working in Windows though.

  • @ChrisMeans
    @ChrisMeans 9 місяців тому

    Thanks!

    • @codetothemoon
      @codetothemoon  9 місяців тому

      wow thank you so much for the support! much appreciated! really happy you got something out of the video.

  • @robbbieraphaelday999
    @robbbieraphaelday999 11 місяців тому +1

    Thank you!

    • @codetothemoon
      @codetothemoon  11 місяців тому

      thanks for watching, glad you got something out of it!

  • @eyemazed
    @eyemazed 11 місяців тому +1

    Awesome tutorial. Would it be possible to make a video on how to use one of these models to further fine train it on our own corpus of content? Sort of make it learn additional things on top, is that too complicated?

  • @KlausSzyska
    @KlausSzyska 10 місяців тому

    your example works nicely in Jupyter notebook after making sure the working version of 'transformers-4.28.0.dev0' is installed.

    • @KlausSzyska
      @KlausSzyska 10 місяців тому

      spoke too soon. Base model works nicely for the [blue] sky but the common pet question gets the spokesman for the Associated Press involved:
      Enter something: What is a common pet for humans to have?
      ["W: I'm sorry, but I'm not sure what happened. I'm not sure what happened."]
      ['h - h - h']
      ['spokesman for the Associated Press said the agency was unable to comment on the matter.']
      ['t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t']
      [' ']
      ['i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i']

    • @KlausSzyska
      @KlausSzyska 10 місяців тому

      Interestingly this bad (t5-large) model behavior only happens when run inside your FOR loop structure. When a FOR loop uses a function call then the large model gets the answers right even for Apollo13 when the question is more specific, i.e. aborted landing.

  • @WillyKusimba
    @WillyKusimba 10 місяців тому +1

    Impressive!

  • @valueray
    @valueray 10 місяців тому

    how can i add my own corpus to a base model and weighting it? i have a big searchable database but want to make it "transformer searchable". Does such a solution exist? Pls someone point me in the right direction. Thank You

  • @PeterAdiSaputro
    @PeterAdiSaputro 9 місяців тому

    The diagram in minute 9:30 reminds me when I tried to learn SVMs.

  • @a_k__
    @a_k__ 11 місяців тому +2

    I love this one

  • @IroAppe
    @IroAppe 11 місяців тому

    I know this tutorial wanted to show how to save on the expenses and run a model on your own machine. But ironically, the outcome actually showed me what we're paying OpenAI for. These models get already very basic things wrong, really. I don't know if it's much of use with these models, yet.

    • @normanlove222
      @normanlove222 11 місяців тому

      the idea is to combine this idea with our local data and then it becomes powerful. we would do that via Llama_index. Just not too many videos out there on how to do this.

  • @animeshsahu2803
    @animeshsahu2803 11 місяців тому +2

    Just a question, do these language models hold context for previous questions and replies?

    • @codetothemoon
      @codetothemoon  11 місяців тому +1

      The way I'm using it in this video - the answer is no. There are models specifically trained for this task though, check this out huggingface.co/models?pipeline_tag=conversational&sort=downloads

    • @animeshsahu2803
      @animeshsahu2803 11 місяців тому

      @@codetothemoon hey thanks, appreciate your time to reply back.

  • @oneproductivemusk1pm565
    @oneproductivemusk1pm565 11 місяців тому

    Isn't this fire 🔥 ship narrator?
    Sounds like him!
    Is this your new channel?

  • @flipper71100
    @flipper71100 10 місяців тому +1

    I wanted to clear up something about the example. When the tokenizer separated the word "undoubtedly" into two tokens, "_un" and "doubtedly," you mentioned it was mainly for efficiency. However, what I learned is that it's called subword tokenization strategy, and it helps the model handle words not found in its training data.
    Is not that the purpose ?

    • @codetothemoon
      @codetothemoon  10 місяців тому +1

      I think that's another aspect of it. But it's also my understanding (could be wrong) that it's also to simplify the model (and potentially improve performance) by having it learn the semantic meaning of the "root" or a word, instead of having to independently learn the semantic meaning of each slightly different flavor of a word

    • @flipper71100
      @flipper71100 10 місяців тому

      @@codetothemoon I dug deep and found something interesting and you were right about the performance point , he is a summary of what I found:
      the large language models typically have a fixed vocabulary size due to memory constraints. Subword tokenization enables the model to represent a more significant number of unique words by encoding them as subword tokens. This helps in handling a wide range of words while keeping the vocabulary manageable.
      Also, another goal is to reduce the overall number of tokens in a text, making the input sequences shorter. This leads to more efficient memory utilization and faster computation during training and inference

  • @being-awe-some
    @being-awe-some Рік тому +1

    pyhton IDE you're showing is within hugging face?

    • @codetothemoon
      @codetothemoon  Рік тому

      actually I'm using doom emacs in this video!

  • @VaibhavShewale
    @VaibhavShewale 10 місяців тому +2

    so to use it do we have to download the model first?

    • @codetothemoon
      @codetothemoon  10 місяців тому

      yes, but this will be done automatically the first time you run the program (if written the way we do here). By default, the Transformers package completely abstracts away the process of downloading the model

  • @s1v7
    @s1v7 9 місяців тому

    actually even scary, how simple everything is

  • @ikemkrueger
    @ikemkrueger 11 місяців тому +1

    How can I fix errors the model makes, like the ones with the Apollo question?

    • @codetothemoon
      @codetothemoon  11 місяців тому +1

      I'm willing to bet it could be fixed with fine tuning, especially if you have some idea of the specific types of questions the model will be expected to answer accurately. The other option is to use larger flavors of the same model. You should be able to get what you're going for using one or both of those strategies

  • @rembrandtharmenszoon5889
    @rembrandtharmenszoon5889 11 місяців тому +1

    Hey you use vim, it's kind of unexpected since the context is AI and stuff. That's great! 👍
    But can I ask why though?

    • @codetothemoon
      @codetothemoon  11 місяців тому

      actually its doom emacs! Which has evil mode (vim bindings) enabled by default. I use the vim motions because I feel like they make me more productive, and I use emacs mostly because of org mode - if there was no org mode I might be using Neovim.

  • @autohmae
    @autohmae 11 місяців тому +5

    I realized something, the tokens is what we also see in The Matrix digital rain ? 🙂

  • @occamsrazor1285
    @occamsrazor1285 11 місяців тому

    13:49 There were two: Apollo 8 and 10

  • @XzcutioneR2
    @XzcutioneR2 Рік тому

    How do I get the max amount of tokens supported by each model?
    I'm trying to build a generic scraper that scrapes the given documentation of an app, finds all relevant REST APIs, webhooks etc along with the relevant details. I would then convert them into embeddings and insert into a database. I'll then search on these semantically.
    What I'm struggling with right now is how to find the structure of the documentation. I think a LLM model would help with that but with most, I feel that I'll have a problem with the max number of tokens when I pass the HTML to them. I have access to GPT-4 but AFAIK, the 32k model hasn't been released yet. And even if I had access to it, I'll run into cost constraints.

  • @theLowestPointInMyLife
    @theLowestPointInMyLife Рік тому +1

    What use are these models to your own app/site? Wouldn't you need to train them on your own data that's relevant to your site?

    • @codetothemoon
      @codetothemoon  Рік тому +14

      They are pretty capable with no fine tuning, it’s very possible they might already do what you need without any additional training. But fine tuning can definitely make them better - I’m up for making a video about that if folks are interested

    • @theLowestPointInMyLife
      @theLowestPointInMyLife Рік тому +1

      @@codetothemoon I'm thinking like site functionality, so a user is asking a question about how to use the site in some way, or some rules about the site. If it's just a general model I don't see the point of every site having one.

    • @codetothemoon
      @codetothemoon  Рік тому +3

      @@theLowestPointInMyLife right - actually you can cover quite a bit of ground by having an app-specific prompts that you interpolate user-specific values into. Using this approach you can generate app-specific output without actually doing any fine tuning to the model itself. That may or may not be sufficient depending on what you're doing.

    • @anuradhaamarasinghe9649
      @anuradhaamarasinghe9649 Рік тому +6

      @@codetothemoon Please create a video about fine-tuning if you can. Thanks for this one. Great stuff as always.

    • @ThatGuyJamal
      @ThatGuyJamal Рік тому +2

      @@codetothemoon I would be interesting in seeing a video on how to train one of these open source models ourselves if possible!

  • @ma34529
    @ma34529 11 місяців тому +1

    I love the random Rust shoutout

    • @codetothemoon
      @codetothemoon  11 місяців тому +1

      not entirely random - much of my audience comes from my Rust content 🦀 gotta build a bit of a bridge there!

    • @ma34529
      @ma34529 11 місяців тому

      @@codetothemoon I'm a rustacean, I loved it.

  • @shanukumarsharma9264
    @shanukumarsharma9264 10 місяців тому +2

    Subscribed. Just include a few practical projects. Most people just want to replicate. How to make useful things which might work in real world....

    • @codetothemoon
      @codetothemoon  10 місяців тому +1

      nice, very happy to have you here! I do plan to make more videos about practical applications of language models. Stay tuned...

  • @snk-js
    @snk-js Рік тому +7

    do you fed the AI with all your channel and now it does spit out the whole video for you? it's super cool those edit tho

    • @codetothemoon
      @codetothemoon  Рік тому +3

      I wish! Definitely spent many hours editing this one 😎

  • @edgeeffect
    @edgeeffect 11 місяців тому

    Disappointed it didn't know about Apollo 10!

  • @jxn22
    @jxn22 11 місяців тому +1

    I: "In whose legion fights the most formidable armed combatant?" O: "british army" 🤦‍♂ Great video, CttM.

  • @plumbing1
    @plumbing1 Рік тому +1

    Interesting

  • @jobilthomas64
    @jobilthomas64 9 місяців тому

    🎉

  • @ToddDunning
    @ToddDunning 10 місяців тому +1

    Typescript pleeze, or is that not cool this week?

    • @codetothemoon
      @codetothemoon  10 місяців тому

      not sure if I'll make it to a TypeScript version, though I imagine such a thing would be immensely popular. I am likely going to make a chatbot webapp in 100% Rust - it might be fun to do a followup of the same webapp using something like SvelteKit - we'll see

  • @itsmedheeraj
    @itsmedheeraj 10 місяців тому +1

    Before this Video, AI was blackbox for me, now i know basics, how it works

    • @codetothemoon
      @codetothemoon  10 місяців тому +1

      nice, really happy you got something out of the video!

  • @yash1152
    @yash1152 11 місяців тому

    2:24 tasks of NLP

  • @hotdog2c
    @hotdog2c 11 місяців тому +3

    Have you found one model to be particularly good for coding?

    • @codetothemoon
      @codetothemoon  11 місяців тому +2

      I haven't experimented with enough open source models to have a good answer to this, sorry!

  • @hubstrangers3450
    @hubstrangers3450 10 місяців тому

    Thank you, however, believe most of the famous tech corps where aware of this space, prior 2017 paper (~7-5 yrs), and was masking this technology from the public space, though they used public dataset, Public should really seek for accountability using a proper legal framework, if only politicians were skillful, knowledgeable and talent enough, least they could seek from these tech corps is to provide cloud platform workspace for closer to a zero pricing scenario.

  • @Jay-kb7if
    @Jay-kb7if 10 місяців тому

    I suspect tokenization stems from cognitive psychology and morphemes. If a token is a meaning unit then a morpheme is the smallest possible meaning unit. I.e., un + happy, un changes the meaning of the word.

    • @codetothemoon
      @codetothemoon  10 місяців тому +1

      I suspect this is true! it's interesting how advances in language models might wind up providing more hints about how the human brain works...

    • @Jay-kb7if
      @Jay-kb7if 10 місяців тому

      @@codetothemoon Yep! I lecture in cognitive psychology and there are a lot more parallels. I did a class with a huggingFace GPT model where you could change the token length to illustrate meaningful sentence production/comprehension (i.e., Wernicke's aphasia). I am fairly certain that an interdisciplinary field of academia will spawn in the next few years that marry machine learning and cognitive psychology. I'm sure there's the odd journal here and there but I think it'll take the place of conventional cognitive psychology.

  • @olsuhvlad
    @olsuhvlad 11 місяців тому +1

    35 But some man will say, How are the dead raised up? and with what body do they come?
    36 Thou fool, that which thou sowest is not quickened, except it die:
    37 And that which thou sowest, thou sowest not that body that shall be, but bare grain, it may chance of wheat, or of some other grain:
    38 But God giveth it a body as it hath pleased him, and to every seed his own body.
    39 All flesh is not the same flesh: but there is one kind of flesh of men, another flesh of beasts, another of fishes, and another of birds.
    40 There are also celestial bodies, and bodies terrestrial: but the glory of the celestial is one, and the glory of the terrestrial is another.
    41 There is one glory of the sun, and another glory of the moon, and another glory of the stars: for one star differeth from another star in glory.
    42 So also is the resurrection of the dead. It is sown in corruption; it is raised in incorruption:
    43 It is sown in dishonour; it is raised in glory: it is sown in weakness; it is raised in power:
    44 It is sown a natural body; it is raised a spiritual body. There is a natural body, and there is a spiritual body.
    45 And so it is written, The first man Adam was made a living soul; the last Adam was made a quickening spirit.
    46 Howbeit that was not first which is spiritual, but that which is natural; and afterward that which is spiritual.
    47 The first man is of the earth, earthy: the second man is the Lord from heaven.
    48 As is the earthy, such are they also that are earthy: and as is the heavenly, such are they also that are heavenly.
    49 And as we have borne the image of the earthy, we shall also bear the image of the heavenly.
    50 Now this I say, brethren, that flesh and blood cannot inherit the kingdom of God; neither doth corruption inherit incorruption.
    51 Behold, I shew you a mystery; We shall not all sleep, but we shall all be changed,
    52 In a moment, in the twinkling of an eye, at the last trump: for the trumpet shall sound, and the dead shall be raised incorruptible, and we shall be changed.
    53 For this corruptible must put on incorruption, and this mortal must put on immortality.
    54 So when this corruptible shall have put on incorruption, and this mortal shall have put on immortality, then shall be brought to pass the saying that is written, Death is swallowed up in victory.
    55 O death, where is thy sting? O grave, where is thy victory?
    56 The sting of death is sin; and the strength of sin is the law.
    57 But thanks be to God, which giveth us the victory through our Lord Jesus Christ.
    (1Co.15:35-57)

    • @codetothemoon
      @codetothemoon  11 місяців тому

      do you think LLMs could be used to create a more accessible translation of the Bible?

  • @dinoscheidt
    @dinoscheidt 11 місяців тому +1

    🚨 Keep in mind, that many open source models DO NOT have commercial licensing

    • @codetothemoon
      @codetothemoon  11 місяців тому

      great point! I think I originally mentioned this in the video but I may have left it out of the final cut. But the one we're using in the video (T5 Flan) is Apache 2.0 and thus should be fine to use for commercial purposes.

    • @dinoscheidt
      @dinoscheidt 11 місяців тому

      @@codetothemoon Yes, the ones you showed are save. Just wanted to point it out… ran in a start-up last week that proudly pitched their product and than were devastated that a superficial tech DD reveled that they break licenses left and right. Licenses always have been the annoying accounting part everyone dreads, but for ML models it’s currently 10x as confusing. And probably the only reason, why the „you can actually run stuff locally“ needs extra care in that department. Anyway: Great video as always Mr Moon !

  • @jereziah
    @jereziah Рік тому +2

    ** and * are called 'splat' operators when they're used this way: ua-cam.com/video/tL1zltXuHO8/v-deo.html

  • @BlackIcexxi
    @BlackIcexxi 11 місяців тому

    what in the hell is that line number schema? Lollll. Great video though!

  • @anon_y_mousse
    @anon_y_mousse 11 місяців тому

    Interesting, but let's see you do it from scratch in C.

  • @alurma
    @alurma Рік тому +4

    Awesome video, thanks, this helps!
    Not a fan of clickbait title though

    • @codetothemoon
      @codetothemoon  Рік тому +1

      Thanks for the feedback! What do you think a better title would be?

    • @burnere633
      @burnere633 Рік тому +2

      @John *flan*-t5

    • @burnere633
      @burnere633 Рік тому +1

      I thought it was less clickbait and more tongue-in-cheek (in retrospect -- I didn't have the context until I watched the video).

    • @codetothemoon
      @codetothemoon  Рік тому +1

      @@John-wh9oh flan is an amazing dessert 🍮!

    • @poulticegeist
      @poulticegeist Рік тому +1

      ​@@John-wh9ohflan is the name of the library primarily used in this video, which also happens to be the name of a dessert

  • @browaruspierogus2182
    @browaruspierogus2182 11 місяців тому +1

    They teach how to import pretrained data but never show how to train models and create system of neural network layers to use it as normal dev would
    all that AI stuff is just monopoly on computation patterns

    • @codetothemoon
      @codetothemoon  11 місяців тому

      pretrained models are an incredible resource, as the vast majority of developers don't have the resources or budget to perform such trainings. I'm not sure what you mean by "create a system of neural network layers to use as a normal dev would", as the models in this video are built upon neural networks

    • @browaruspierogus2182
      @browaruspierogus2182 11 місяців тому

      @@codetothemoon to show how to build NN matrixes with predictors for specific models like shape recognition or sound recognition

  • @mail2toan
    @mail2toan 10 місяців тому

    I took it a little too far. It's totally me. 😂

  • @DrCHE42
    @DrCHE42 9 місяців тому +1

    Great presentation, but the presenter needs, IMHO, to speak more slowly

    • @codetothemoon
      @codetothemoon  9 місяців тому

      thanks and thanks for the feedback!

  • @astroganov
    @astroganov 11 місяців тому +1

    Thanks for the video 👍.
    I just can't stop wondering, how come such an impossible poor, slow and inconvenient programming language with the development environment from the past century became a standard for machine learning tasks... It doesn't even have method names autocomplete, it doesn't show the full list of object's methods available, no parameters list, no types - just nothing. 💩 Crazy. return_tensors="pt" 🤦

  • @TheOleHermit
    @TheOleHermit 10 місяців тому

    Been self learning Arduino & Python on YT for several projects over the past few years. But, I have absolutely no idea what is being said in this video. Must be a different language than English. 🤷‍♂
    I'll circle back around in another couple of years, after gaining some more smarts.

  • @robertlawson4295
    @robertlawson4295 11 місяців тому +2

    Just a bit of food for thought ... all CPU chips ONLY understand how to manipulate bits and bytes in registers. EVERY computer language is simply a way to create those bit patterns to feed the hardware chip. Each language chooses how much abstraction to use and how the abstractions are interpreted. Some do it this way and some do it that way, but the end result in ALL cases is a sequence of machine (binary) codes that the Assembly language for the chip can write directly with NO abstractions except, perhaps, some mnemonics (memory aid) to avoid memorizing all the binary values associated with each machine instruction. Programming, paradoxically, is much easier in Assembly language than all others because you've gotten rid of all abstractions and interpretations that you have to wrap your head around. Of course, you still have the abstractions needed for solving the problem you are tackling but you don't have abstractions within abstractions. That's what makes computer programming challenging. Just sayin' ... 😉

    • @codetothemoon
      @codetothemoon  11 місяців тому +4

      I've never encountered a problem that was easier to solve in assembly than a higher level language. The abstractions of higher level languages are there to reduce our cognitive load, and they are incredibly effective at doing so

    • @robertlawson4295
      @robertlawson4295 11 місяців тому +1

      @@codetothemoon Absolutely, I agree. Abstracting things is important in many fields, not just programming. I was principally interested in peeling the onion to show why something like programming CAN be made more difficult because of abstractions but, of course, we have to understand that creating abstractions is an art in itself. It can be done well or done poorly. The effectiveness of using assembly language also depends on the programmer having a rock-solid understanding of the underlying electronics hardware. As an electronics engineer, I have had to work with many chips that didn't have compilers available so all coding had to be done in assembly and carefully mapped out with only the most rudimentary abstractions. This is also true of the microcoding that exists internally in every chip to interpret the binary machine code instructions themselves.

  • @sabin97
    @sabin97 11 місяців тому +1

    just came here for the thumbnail.
    we're not "coders" a trained monkey can code.
    we are programmers.
    if you dont understand the difference, it's the same difference between writing and typing.
    yeah......typing/coding is just hitting the keys in order to form the words. writing/programming is thinking.
    not even gonna watch the video. the thumnail earned it a dislike.