Evaluating LLM-based Applications
Вставка
- Опубліковано 2 жов 2024
- Evaluating LLM-based applications can feel like more of an art than a science. In this workshop, we'll give a hands-on introduction to evaluating language models. You'll come away with knowledge and tools you can use to evaluate your own applications, and answers to questions like:
Where do I get evaluation data from, anyway?
Is it possible to evaluate generative models in an automated way?
What metrics can I use?
What's the role of human evaluation?
Talk by: Josh Tobin
Here’s more to explore:
LLM Compact Guide: dbricks.co/43W... Big Book of MLOps: dbricks.co/3r0...
Connect with us: Website: databricks.com
Twitter: / databricks
LinkedIn: / databricks
Instagram: / databricksinc
Facebook: / databricksinc
This is an exellent coverage of the challenging task of llm evaluatuon
Evaluations aside, I really enjoyed the presentation. I was hooked. Great story-telling skills Josh. Thanks for sharing your experience. We count on volunteers like you to spread knowledge.
54127 George Stream
So happy to see someone summarizing the difficulty of evaluation of LLM applications so well
54289 Cristopher Glen
"Your opininon on LLMs does not matter" - I found this to be a great quote
Thank you for clearing this concepts. Would like to see more videos from you on evaluation frameworks, methods.
Good One!
Good work
74728 Williamson Hills
Very generic, expected something more tangible! Sounds common sense which might work or might not work
Just what I was after. Thanks
Great talk, thanks for sharing
Great storyline