Overview
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a novel approach to detecting errors in language model outputs through "diversity measures" in this 12-minute video. Learn about domain-independent proxies for failure in language model queries, addressing challenges like hallucination. Discover how these measures can be used to assess uncertainty in language model results. Gain insights into the key ideas, self-consistency, and results of this research. Access the preprint and source code for further study. Delve into the exciting intersection of symbolic methods and deep learning, with content derived from an AI course at Arizona State University.
Syllabus
Intro
Key Idea
Selfconsistency
Results
Taught by
Neuro Symbolic