Overview
Learn about Knowledge-Adapted Fine-Tuning (KnowAda), a groundbreaking approach to reduce hallucinations in vision-language models (VLMs) through this 26-minute educational video. Explore how smaller VLMs with up to 7B parameters struggle with balancing descriptive richness and factual accuracy in medical imaging applications. Discover insights from John Hopkins University researchers on probing VLM knowledge through generated visual questions and implementing automatic caption adaptation. Understand the Decomposed NLI (DNLI) evaluation framework that breaks down captions into atomic claims for precise accuracy assessment. Examine the limitations of current medical VLMs, the importance of pre-training in fine-tuning processes, and the anticipated developments in new pre-trained models. Gain practical knowledge about implementing KnowAda methodology, which has demonstrated superior performance in maintaining high descriptiveness while significantly reducing contradiction rates compared to traditional methods.
Syllabus
VLM 8B in MedAI
Identify the problem
Insights by John Hopkins Univ
Fine-Tuning depends on Pre-Training
Fine-Tune Medical VLMs
Knowledge Adapt KnowAda
Limit of medical VLMs
New Solutions
We need new Pre-Trained Models
Delay of new VLMs OpenAI
Taught by
Discover AI