Bert Score for Contextual Similarity for RAG Evaluation

Поділитися
Вставка
  • Опубліковано 14 січ 2025

КОМЕНТАРІ • 13

  • @manishsharma2211
    @manishsharma2211 Рік тому +2

    Quick Correction : ROUGE is used for summarization and BLEU is used for translation primarily [ time stamp : 4:04 ]

  • @SonGoku-pc7jl
    @SonGoku-pc7jl Рік тому +1

    thanks for all! :) tomorrow i will watch your others videos of evaluate llm and rag :)

  • @gotitgotya
    @gotitgotya 5 місяців тому

    great work man....thank you so much for uploading such informative videos❤❤

  • @VenkatesanVenkat-fd4hg
    @VenkatesanVenkat-fd4hg Рік тому

    Thanks for valuable video as usual. Waiting for multimodal and unstructured files/applying RAG

  • @soumilyade1057
    @soumilyade1057 9 місяців тому

    The library that you have used mentions use of custom models in point 3 of ReadMe. But, there's no parameter by the name "model" or "num_layers". I was wondering if you have figured out what's going on there.

  • @ShreyaSingh-wp9yk
    @ShreyaSingh-wp9yk Рік тому

    thanks for uploading the video. one quick question how it is different from rogue, bleu and meteor, as they are also recall and precision-based.can we use rogue, meteor and bertscore if I am evaluating chatbot and why. Please excuse if this ques sounds naive, I am very much new to this

  • @livelaughmotivate94
    @livelaughmotivate94 10 місяців тому

    For Text Generation RAG, Bert Score won't work?

  • @ArpitBhavsar-z2c
    @ArpitBhavsar-z2c Рік тому +1

    Though, in the second example, it gave 85% similarity
    Weird right?

    • @CibeSridharanK
      @CibeSridharanK 8 місяців тому +2

      exactly thats why his reaction is strange

    • @encianhoratiu5301
      @encianhoratiu5301 3 місяці тому

      yeah, not so good to evalute a llm generation... I don't blame him perahps there isn't a better way.

  • @suryaaonline1935
    @suryaaonline1935 19 днів тому

    Thanks Man!