Stealing Part of a Production LLM | API protects LLMs no more

Поділитися
Вставка
  • Опубліковано 10 гру 2024

КОМЕНТАРІ • 93

  • @MyrLin8
    @MyrLin8 6 місяців тому +4

    Excellent! White paper from LLM monitoring, simply excellent :) ... off to the reading library. thanks muchly.

  • @jmirodg7094
    @jmirodg7094 7 місяців тому +4

    Thanks! Concentrate on your PhD writing it is important! We can wait for the next video.

  • @robthomas2049
    @robthomas2049 8 місяців тому +10

    Can't say I really understand most of what they have done but it seems totally wild and real creative how they managed to work it out. Good luck with the thesis :D

  • @HunterMayer
    @HunterMayer 8 місяців тому +10

    Wow. I Lived. I have a long way to go to understand it all, but I held on!

  • @HunterMayer
    @HunterMayer 8 місяців тому +7

    Good luck with the paper, remember to take some time for yourself afterwards! I look forward to your next contribution.

  • @ricowallaby
    @ricowallaby 8 місяців тому +5

    Wow. . . I'm glad I've found your channel very informative, I'm still trying to wrap my head around those articles, love your accent and your smile, looking forward to more videos, best of luck in your thesis, cheers from Sydney.

  • @praxis22
    @praxis22 8 місяців тому +7

    Good luck defending your thesis, your accent is one of the reasons i watch.

  • @IndoorAdventurer1996
    @IndoorAdventurer1996 8 місяців тому +16

    Really cool video. Good luck with your thesis! ^^

  • @bernieapodaca2912
    @bernieapodaca2912 8 місяців тому +7

    WOW! 😅 I watched the whole video! I think my brain was somewhat cramping, but I grasped the concept. Good job and good luck with your thesis!!!! I subscribed but I’ll wait for my brain to heal a bit before I jump into another great video 🙃🙏👍

  • @MaJetiGizzle
    @MaJetiGizzle 8 місяців тому +4

    For what it’s worth, I completely understand your accent (it’s not very heavy in English), and wow I was completely overthinking it with my guesses on the paper on Patreon. 😅
    I’m glad I can now steal all the logits if I wanted! 😜
    Another fantastic paper breakdown as always!

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +2

      Thank you! 😃 Your guesses on Patreon where highly informative. I told you I was humbled about how many good papers I did not know. 🤭

  • @visheshmittal468
    @visheshmittal468 8 місяців тому +11

    Hey amazing content, quick note 2nd paper is from University of southern California (My uni) so just pointing that out!

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +4

      Thanks! I've butchered it at 3:15, I did not notice that it was not "California" that came out of my mouth, sorry. 🙈

  • @nathanbanks2354
    @nathanbanks2354 8 місяців тому +6

    Good luck with your thesis! If you want try to do a summary video of your thesis after you submit it and before your defense, I'd watch it. Could be good practice for the defense too. I'm sure you know more about the topic than anyone else in the world...new research is cool.

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +1

      Thanks, I think I will do something like that. :)

  • @williamchamberlain2263
    @williamchamberlain2263 8 місяців тому +7

    Awesome. Good video editing snd pacing too.

  • @BooleanDisorder
    @BooleanDisorder 8 місяців тому +4

    So cool. Thanks mate! I wonder how other architectures would fare against this attack.
    Also, super good luck on your thesis mate!!! I would love to see what you've been researching!

    • @AICoffeeBreak
      @AICoffeeBreak  7 місяців тому +1

      I think quite badly as long as they have that hidden-dim to logits mapping. And most classification models have that. 😅

  • @marcfruchtman9473
    @marcfruchtman9473 8 місяців тому +7

    I am not a fan of the Title of the paper. First of all, analyzing data to determine a hidden aspect of the data where such data is publicly available is not stealing. And it does a huge disservice to professionals in the industry to call it that. It would serve well as a Joke title... but no professional in the information world should be using those kinds of titles.
    Other than that, my hat goes off to the researchers for analyzing the data and discovering ways to reveal this information.
    Thank you for making this video... Some of the best analysis on the net!

    • @Acceleratedpayloads
      @Acceleratedpayloads 8 місяців тому +4

      As researchers who take an adversarial approach to security, we sometimes must use prescriptive language when discussing methodologies of morally reprehensible acts. It's called red teaming.
      How else do you propose to succinctly describe such an action?

    • @marcfruchtman9473
      @marcfruchtman9473 8 місяців тому +3

      ​@@Acceleratedpayloads "Speculative API Analysis: Statistically Inferring Unpublished LLM Parameters and Updates". You could also add in "Penetration Testing" as part of the title to provide the information that this was done with security in mind: *"API Penetration Testing: Statistically Inferring Unpublished LLM Parameters and Updates"*
      I realize that flashy titles are the new "in" thing... but I really really don't want to see: "STUNNING new technique steals LLM secrets in SHOCKING new API HACK, GPT loses Everything!" hehehe

    • @MDNQ-ud1ty
      @MDNQ-ud1ty 8 місяців тому

      Our entire world is based on theft. Theft of human energy that is focused on the things those that do the steal. It will only get worse as capitalism becomes more late stage and the best capitalists start to more in to pharaohs. Happy pyramid building.

    • @CVE42287
      @CVE42287 8 місяців тому +2

      ​@@AcceleratedpayloadsYou don't sound like someone that's very good at it tbh

    • @fatherfoxstrongpaw8968
      @fatherfoxstrongpaw8968 8 місяців тому

      @@Acceleratedpayloads I have a suggestion. how about honestly? or perhaps "in context"? just a thought.

  • @DeGandalf
    @DeGandalf 8 місяців тому +11

    Good luck with your PhD! While I wasn't missing your videos (as in I prefer higher quality videos instead of regular ones, which you are nailing), I'll definitely be looking forward to them.

  • @AmCanTech
    @AmCanTech 7 місяців тому +4

    So much linear algebra is involved in LLMs

    • @AICoffeeBreak
      @AICoffeeBreak  7 місяців тому +1

      They are a pile of linear algebra.

  • @gregsLyrics
    @gregsLyrics 7 місяців тому +2

    The intellect of a woman has always melted my heart. I think I just fell in love with Letitia. That is the most beautiful lecture I have ever enjoyed. ❤

  • @supernewuser
    @supernewuser 8 місяців тому +3

    great content, glad I found the channel

  • @jackpisso1761
    @jackpisso1761 8 місяців тому +4

    I really like your content. Thank you so much for making it!

  • @Thomas-gk42
    @Thomas-gk42 8 місяців тому +8

    Just here to say, your accent is wonderful, romanian I guess from your name. All the best.

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +3

      Exactly, the ending "escu" is a perfect giveaway.

  • @a0z9
    @a0z9 8 місяців тому +4

    Pronuncias muy bien. Es muy fácil hachear la ia explotando probabilisticamente los sesgos que le introducen para limitar su potencia.

  • @cosmic_reef_17
    @cosmic_reef_17 8 місяців тому +5

    Perfect way to start a week with one of your videos. Great work!

  • @glymjerlackless8605
    @glymjerlackless8605 8 місяців тому +9

    The idea that GPT-3.5-Turbo is just a 7b parameter model seems a little bit unlikely to me, its performance is leagues beyond the open-source 7b models that at least I'm familiar with. If they hadn't explicitly addressed it in the paper I would have assumed that it was a 8x7B or similar model. At least the performance of gpt-turbo-3.5 and mixtral is roughly the same.

    • @tsomerville1970
      @tsomerville1970 7 місяців тому +1

      OpenChat3.5 is 7B (mistral variant) spans what 3.5gpt does. So I believe it’s possible.
      However, I believe you could be right in that there could be MoE as well - or special teacher / student training that openAI did to help 3.5

  • @MachineLearningStreetTalk
    @MachineLearningStreetTalk 8 місяців тому +4

    Here we go!!!

  • @DaNa-px7ol
    @DaNa-px7ol 8 місяців тому +5

    Good luck with the Thesis!

  • @drummatick
    @drummatick 8 місяців тому +4

    As "d" is the hidden embedding dimension, is it guaranteed somehow that the logits and embeddings themselves lie in d-dimensional space? Or they probably lie in lower dimensional sub-space?

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому

      Great observation. It could happen that they lie in a smaller-than-d dimensional space. They mention this in the Finlayson et al. Paper in a footnote. :)

  • @MyrLin8
    @MyrLin8 6 місяців тому +4

    LOL yup, still here ;)

  • @AbdallahAbdallah
    @AbdallahAbdallah 2 місяці тому +2

    What software and video editing tools do you use for creating this great content ?

    • @AICoffeeBreak
      @AICoffeeBreak  2 місяці тому +2

      I make all visuals (including the drawings in PowerPoint. 😅). I use Adobe Premiere Pro for editing (this is also the stage Ms. Coffee Bean comes into the picture).

    • @AbdallahAbdallah
      @AbdallahAbdallah 2 місяці тому +1

      @@AICoffeeBreak this is so impressive. I couldn't even imagine that you can do that much visual using PowerPoint. You must be a guru level in power point. Maybe this is the topic of another video to make :)

    • @AICoffeeBreak
      @AICoffeeBreak  2 місяці тому

      @AbdallahAbdallah

  • @NatColley-t4z
    @NatColley-t4z 8 місяців тому +1

    Aside from the embarrassment and possible competitive disadvantage, do these papers also imply security risks? Rewriting the model in some way, like an old fashioned code injection? Are these methods available against any/all APIs?

  • @joelbelafa
    @joelbelafa 2 місяці тому +1

    Yep still there. I assume a MoE would make one’s life a bit difficult on hacking Token bias feature. Correct ?

  • @opusdei1151
    @opusdei1151 7 місяців тому +2

    LLM-hacker for the rise xD

  • @AM-yk5yd
    @AM-yk5yd 8 місяців тому +1

    * easiest way to prevent attack is to limit biases. Nobody needs to bias lots of tokens
    * maybe fuse last mlp with embed_out? Right now transformers do dim_embed -> dim_ffn -> dim_embed -> n_vocab. If last mlp outputs n_vocab directly instead of dim_embed, then itll be harder to figure out true dim_embed as mlp already uses non linearity

  • @mohammadmira1626
    @mohammadmira1626 8 місяців тому +4

    Thank you❤

    • @AICoffeeBreak
      @AICoffeeBreak  7 місяців тому

      Thank you for your visit! Hope to see you again!

  • @spencerfunk6697
    @spencerfunk6697 8 місяців тому +4

    my idea is openai is trying to do something similar to apple, as in whenever apple drops a new phone they make all the old ones worse. i think thats a big reason theyre so desperate to keep everything hidden

  • @ScottzPlaylists
    @ScottzPlaylists 8 місяців тому +5

    @AICoffeeBreak
    The paper Reads "University of Southern California", but ...
    You said "University of Southern Carolina" first time, and "University of South Carolina" 2nd time.
    -- I'm sorry, my INTJ personality picks up on the little details. 😄

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +2

      Thanks for pointing out, and again: sorry, I had a brain fart there. 😅

    • @ScottzPlaylists
      @ScottzPlaylists 8 місяців тому +2

      @@AICoffeeBreak Must have been Coffee Beans on the Brain...🤣

  • @juanmanuelcirotorres6155
    @juanmanuelcirotorres6155 8 місяців тому +4

    thank you

  • @Jamaleum
    @Jamaleum 8 місяців тому +5

    3:15 *California :D

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +3

      Oh, yes! Sorry! 🙈

    • @ScottzPlaylists
      @ScottzPlaylists 8 місяців тому +2

      I see someone else caught that now that I looked ... Is your personality INTJ also?

  • @bmebri1
    @bmebri1 8 місяців тому +3

    💋🦒

  • @daspradeep
    @daspradeep 8 місяців тому +4

    not a big deal and it's mentioned in the paper as well

  • @jasdeepsinghgrover2470
    @jasdeepsinghgrover2470 7 місяців тому

    Next step... Get the model using distillation... And obviously it will not be published..😂😂

  • @Kuratius
    @Kuratius 7 місяців тому

    If this approach works for LLMs in general, I wouldn't be suprised if some variation of it worked on human brains. That would be way crazier.

  • @NoNameAtAll2
    @NoNameAtAll2 8 місяців тому

    what happens when you apply this to a human?

  • @fatherfoxstrongpaw8968
    @fatherfoxstrongpaw8968 8 місяців тому

    clickbait! fake news! i call BS on the title. *LeSigh...* This is like saying a smart kid figuring out a magic trick is stealing the magician's magic. quit misrepresenting to get views, it cheapens and invalidates your good work!!!

    • @DerPylz
      @DerPylz 8 місяців тому +2

      This is literally the title of the paper that is being explained here... There is no misrepresentation to get views here.

    • @fatherfoxstrongpaw8968
      @fatherfoxstrongpaw8968 8 місяців тому

      @@DerPylz i'm sorry, i missed that. could you point that out for me? (timestamp?)

    • @DerPylz
      @DerPylz 8 місяців тому +3

      @@fatherfoxstrongpaw8968 The first time the papers are shown is at 0:35, then again at 1:45 and a few more times throughout the video. The papers are also linked in the description, if you want to have a closer look yourself.

    • @fatherfoxstrongpaw8968
      @fatherfoxstrongpaw8968 8 місяців тому

      @@DerPylz yeeaaa... still calling BS. if you can guess or "infer" what it's doing, it's not stealing. sorry.

    • @DerPylz
      @DerPylz 8 місяців тому +3

      @@fatherfoxstrongpaw8968 Wait, is your criticism with the paper or with the video? Because the video accurately explains what the authors of "Stealing Part of a Production Language Model" did in their paper. Of course the video will be titled according to the paper it explains, so that people looking for a explanation of that paper can find the video. So I don't see any false advertising or misrepresentation there. If you don't agree with the paper's title then that's a different discussion, and you might want to take up your criticism with the authors. But your comment saying "it cheapens and invalidates your good work" makes it sound like you're blaming Letitia.