Prompt Injections - An Introduction

Поділитися
Вставка
  • Опубліковано 7 чер 2024
  • Many courses teach prompt engineering and currently pretty much all examples are vulnerable to Prompt Injections. Especially Indirect Prompt Injections are dangerous. They allow untrusted data to take control of the LLM (large language model) and give an AI a new instructions, mission and objective.
    This video aims to raise awareness of this rising problem.
    Injections Lab: colab.research.google.com/dri...
    Prompt Engineering Overview 0:00
    Prompt Injections Explained 2:05
    Indirect Prompt Injection and Examples 4:03
    GPT 3.5 Turbot vs GPT-4 5:55
    Examples of payloads 6:15
    Indirect Injections, Plugins and Tools 8:20
    Algorithmic Adversarial Prompt Creation 10:35
    AI Injections Tutorials + Lab 12:22
    Defenses 12:39
    Thanks 14:40
  • Наука та технологія

КОМЕНТАРІ • 4

  • @ninosawas3568
    @ninosawas3568 5 місяців тому +1

    Great video! Very informative. Interesting to see how the LLMs ability to "pay attention" is such a large exploit. I wonder if mitigating this issue would lead to LLMs being overall less effective at following user instructions

    • @embracethered
      @embracethered  5 місяців тому +1

      Thanks for watching! I believe you are correct, it's a double edged sword. The best mitigation at the moment is to not trust the responses. Unfortunately it's hence impossible at the moment to build a rather generic autonomous agent that uses tools automatically. It's a real bummer, because i think most of us want secure and safe agents.

  • @halfoflemon
    @halfoflemon Рік тому +1

    How about giving it a secret word that should be typed in order to unlock control, like a password? Do you think it will work? Also, does lowering the temperature reduces the chance of successful injection attack?

    • @embracethered
      @embracethered  Рік тому

      Yes, something like that works. I have done it with image models in the past, basically train the model to respond in particular way once a certain object is present. You can check out this blog post on what is possible: embracethered.com/blog/posts/2020/husky-ai-machine-learning-backdoor-model/
      Higher temperature means more "creativity", so it is probably more likely to come up with responses that could be considered insecure, but also less deterministic.