Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Knowledge-Adapted Fine-Tuning for Medical Vision Language Models

Discover AI via YouTube

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Learn about Knowledge-Adapted Fine-Tuning (KnowAda), a groundbreaking approach to reduce hallucinations in vision-language models (VLMs) through this 26-minute educational video. Explore how smaller VLMs with up to 7B parameters struggle with balancing descriptive richness and factual accuracy in medical imaging applications. Discover insights from John Hopkins University researchers on probing VLM knowledge through generated visual questions and implementing automatic caption adaptation. Understand the Decomposed NLI (DNLI) evaluation framework that breaks down captions into atomic claims for precise accuracy assessment. Examine the limitations of current medical VLMs, the importance of pre-training in fine-tuning processes, and the anticipated developments in new pre-trained models. Gain practical knowledge about implementing KnowAda methodology, which has demonstrated superior performance in maintaining high descriptiveness while significantly reducing contradiction rates compared to traditional methods.

Syllabus

VLM 8B in MedAI
Identify the problem
Insights by John Hopkins Univ
Fine-Tuning depends on Pre-Training
Fine-Tune Medical VLMs
Knowledge Adapt KnowAda
Limit of medical VLMs
New Solutions
We need new Pre-Trained Models
Delay of new VLMs OpenAI

Taught by

Discover AI

Reviews

Start your review of Knowledge-Adapted Fine-Tuning for Medical Vision Language Models

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.