Overview
Syllabus
Introduction
Technical Design Flowchart
Installing Chrome
Chromedriver not available and how to fix it
Testing selenium webdriver
Autogen code overview
AI Agents more in-depth
Fetch image overview
Accessing page source
Gotcha with page source
Renaming low resolution link to highest resolution link
Testing the fetch_images script
Revisiting the Autogen code
Autogen in action
Checking the downloaded images
Organizing the images
Using Topaz Gigapixel AI to upscale images
Loading LLM framework overview
Installing text-generation-webui
Showing git hash for text-generation-webui
Downloading llava-v1.5-13b-GPTQ
Support LLaVA v1.5 pull request
Commit log for LLaVA v1.5
Original LLaVA v1.5-13b repository
Possible ? to load llava-v1.5-13b using --load-in-4bit flag in the readme
Downloading the model through CLI
Model Placement in text-generation-webui directory
Multimodal documentation for starting up API
Command to start the server
text-generation-webui GUI
Looking at pull request to see suggested settings
Changing presets in text-generation-webui
Initial trials in the GUI
Comparing concise and verbose prompt instruction
Testing out the text-generation-webui API
Get the IP address of windows from inside Linux
Finding the endpoint/API examples
Testing the API request
Comparing results between the API and the GUI
llava v1.5-13b responding in another language, hallucination ?
Using replicate's original llava v1.5-13b model
Bringing up concise vs. verbose prompt again
Setting up replicate API key locally
Setting up python call to replicate
Running iterate replicate code
Download llava-v1.5-7b model
Setting up llama.cpp framework
Adding models to llama.cpp
Showing llama.cpp commit hash
Starting up llama.cpp server
llama.cpp GUI
llama.cpp API code overview
llama.cpp server API documentation
Taught by
kasukanra