Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Visual Language Models for Edge AI 2.0 - Multi-image Reasoning and In-context Learning

EDGE AI FOUNDATION via YouTube

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Watch a 37-minute conference talk exploring cutting-edge innovations in edge AI technology, where Song Han presents groundbreaking developments in visual language models and model efficiency. Dive into VILA (CVPR'24), a sophisticated visual language model capable of multi-image reasoning and in-context learning, specifically designed for deployment on Jetson Orin Nano. Learn about AWQ (MLSys'24), an advanced 4-bit LLM quantization algorithm that enhances model efficiency, and discover TinyChat, an inference library powering visual language model inference. Understand how these three technologies combine to enable advanced visual reasoning capabilities on edge devices, opening new possibilities for edge AI applications.

Syllabus

GenAI on the Edge Forum - Song Han: Visual Language Models for Edge AI 2.0

Taught by

EDGE AI FOUNDATION

Reviews

Start your review of Visual Language Models for Edge AI 2.0 - Multi-image Reasoning and In-context Learning

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.