Improving LLM Performance Through Evaluation and Few-Shot Examples - Part 2

Improving LLM Performance Through Evaluation and Few-Shot Examples - Part 2

Trelis Research via YouTube Direct link

- Conclusion and future topics

21 of 21

21 of 21

- Conclusion and future topics

Class Central Classrooms beta

YouTube videos curated by Class Central.

Classroom Contents

Improving LLM Performance Through Evaluation and Few-Shot Examples - Part 2

Automatically move to the next video in the Classroom when playback concludes

  1. 1 - Introduction to LLM evals and their importance
  2. 2 - Overview of creating high-quality prompts from scratch
  3. 3 - Emphasis on importance of high-quality examples
  4. 4 - Introduction to systematic approach for creating examples
  5. 5 - Overview of demonstration using touch rugby example
  6. 6 - Introduction to LLM eval repo and UI demonstration
  7. 7 - Start of UI demonstration with pipeline creation
  8. 8 - Creating initial pipeline with Claude Sonnet
  9. 9 - Creating first dataset for touch rugby Q&A
  10. 10 - Setting up evaluation criteria and format requirements
  11. 11 - Demonstration of generating ground truth answers
  12. 12 - Creating second evaluation task
  13. 13 - Introduction to creating few-shot examples
  14. 14 - Setting up pipeline with few-shot examples
  15. 15 - Creating training examples for few-shot learning
  16. 16 - Demonstration of improved performance with few-shot examples
  17. 17 - Discussion of pipeline customization options
  18. 18 - Final tips on judges and evaluation
  19. 19 - Recommendations for managing examples
  20. 20 - Discussion of OpenAI's o1 model considerations
  21. 21 - Conclusion and future topics

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.