Hacking Google Bard: Prompt Injection to Data Exfiltration via Image Markdown Rendering (Demo Video)

Поділитися
Вставка
  • Опубліковано 2 лис 2023
  • Demo video of end to end data exfiltration exploit via a malicious Google Doc. The exploit leverages an indirect prompt injection which injects an image markdown element which is the exfiltration channel.
    This vulnerability was responsibly disclosed to Google VRP on September, 19th 2023 and Google reported it as fixed October, 19th 2023.
    Details in this blog post:
    embracethered.com/blog/posts/...
    Finders and Reporters: @wunderwuzzi23 @rez0__ @KGreshake
  • Наука та технологія

КОМЕНТАРІ • 9

  • @balonikowaty
    @balonikowaty 7 місяців тому +3

    Great work Johann, as always! The more we give access to other data sources. which include documents, the more we expose each other to indirect injection attacks. It is worth pointing out that instructions could have been made in white ink size 0.1, making the document look normal!

  • @fire17102
    @fire17102 7 місяців тому +3

    Read the post, really good
    I guess these sort of procedures will work across many different stacks and companies
    Also I wonder if you log your attempts, probably allot of wisdom can be drawn from your first attempt evolving to the last. You got it on the 10th try. Maybe showing a smart llm all 10 of those could find patterns. Effectively creating a prompt optimizer thay bring you faster results next time.
    All the best

    • @embracethered
      @embracethered  7 місяців тому +1

      Thanks for the note! Yes, this is a very common flaw across LLM apps. Check out some of my other posts about Bing Chat, ChatGPT or Claude.
      Yep, on the iteration count - spot on. A lot of initial tests were around basic validation that injection and reading of chat history worked, then the addition of Image rendering, then in context learning examples to increase reliability of the exploit.

  • @6cylbmw
    @6cylbmw 3 місяці тому +1

    I didn't really understand the vulnerability impact. You are exfiltrating own chat (user A) to own drive (user A) drive. How is it exploitable?

    • @embracethered
      @embracethered  3 місяці тому

      Attacker is causing the Chatbot to send past chat data to attackers server (in this case a google doc is capturing the exfiltrated data).
      Check out the linked blog post, explains it in detail.

  • @petraat8806
    @petraat8806 4 місяці тому

    im trying to understand what just happened please can someone explain

    • @embracethered
      @embracethered  4 місяці тому

      You can read up on the details here: embracethered.com/blog/posts/2023/google-bard-data-exfiltration/
      And if you want to understand the big picture around LLM prompt injections check out this talk m.ua-cam.com/video/qyTSOSDEC5M/v-deo.html
      Thanks for watching!