Multi-Armed Bandits and A/B Testing

Поділитися
Вставка
  • Опубліковано 1 лип 2024
  • Today I'm talking to Sandeep, a PhD student studying Information and Decision Sciences at the University of Minnesota. We talk Multi-Armed Bandits, A/B Testing, and the key differences between the two.
    Check out Sandeep's website: sandeepgangarapu.com/
    Want to be featured in the next mock interview video? Apply here: airtable.com/shrdQrwKK7xxGLm6l
    👉 Subscribe to my data science channel: bit.ly/2xYkyUM
    Use the code "datasciencejay" and get 10% off data science interview prep 🔥 : www.interviewquery.com/pricin...
    ❓ Check out our data science courses: www.interviewquery.com/course...
    🔑 Get professional coaching here: www.interviewquery.com/coachi...
    🐦 Follow us on Twitter: / interview_query
    More from Jay:
    Read my personal blog: datastream.substack.com/
    Follow me on Linkedin: / jay-feng-ab66b049
    Find me on Twitter: / datasciencejay
  • Наука та технологія

КОМЕНТАРІ • 10

  • @tinawang1291
    @tinawang1291 2 роки тому +3

    Learnt something today , thanks! I think for the last example of unlearnai, they will still need to test few real people with placebo to validate their model performance. With a proven working model, they can test mainly with real drug for side effect, etc

  • @YaminiKurra
    @YaminiKurra 2 роки тому +2

    Such a great talk sandy! So proud of you

  • @CruiserPup
    @CruiserPup 2 роки тому

    Wow, this was such a great convo! Thanks Sandeep for sharing your wisdom, going to be checking out your other work!

  • @adhithyajoe1417
    @adhithyajoe1417 2 роки тому +2

    Great content!!

  • @shankars4384
    @shankars4384 9 місяців тому

    This was a great video!

  • @ravennsiregar
    @ravennsiregar 6 місяців тому

    Hello Sandeep, thank you for the quick overrun. Do you mind to tell us how to connect or discuss with you after this session?
    Follow up, so I feel that Multi Armed Bandit is sort of Optimisation Problem given such constraint that it is quite hard and ineffective to perform AB Testing? Do you agree with such motion? Let me know your inputs

  • @sriharshamadala4656
    @sriharshamadala4656 2 роки тому +2

    Its not often you hear a researcher give a high level talk that regular folks can understand. Great talk. Enjoyed it thoroughly. About that 20$ though, whats the algo haha

    • @ravennsiregar
      @ravennsiregar 6 місяців тому +1

      at the moment it is often using UCB/Upper Confidence Bound to maximise utility return. But the overall problem is, in casino the reward is not simply one state. It is far complex than simple one state bandit context tho. The casino example is a mere oversimplifying.

  • @iancheung3587
    @iancheung3587 2 роки тому +1

    What's Sandeep's full name/ linkedin

  • @radio-controlledcouk
    @radio-controlledcouk Рік тому

    You cant use Multi armed bandits in online experimentation because they cause return user bias. MAB's can only be used once per user. The problem is that bandit machines have a fixed probability of payout.... whilst a user of a websites probability of buying something increases over time. This means that if they are switched into a new variation that new variation is more likely to incur an outcome of a sale...... flawed experiment!