Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

Udemy

Generative AI Architectures with LLM, Prompt, RAG, Vector DB

via Udemy

Overview

Design and Integrate AI-Powered S/LLMs into Enterprise Apps using Prompt Engineering, RAG, Fine-Tuning and Vector DBs

What you'll learn:
  • Generative AI Model Architectures (Types of Generative AI Models)
  • Transformer Architecture: Attention is All you Need
  • Large Language Models (LLMs) Architectures
  • Text Generation, Summarization, Q&A, Classification, Sentiment Analysis, Embedding Semantic Search
  • Generate Text with ChatGPT: Understand Capabilities and Limitations of LLMs (Hands-on)
  • Function Calling and Structured Outputs in Large Language Models (LLMs)
  • LLM Providers: OpenAI, Meta AI, Anthropic, Hugging Face, Microsoft, Google and Mistral AI
  • LLM Models: OpenAI ChatGPT, Meta Llama, Anthropic Claude, Google Gemini, Mistral Mixral, xAI Grok
  • SLM Models: OpenAI ChatGPT 4o mini, Meta Llama 3.2 mini, Google Gemma, Microsoft Phi 3.5
  • How to Choose LLM Models: Quality, Speed, Price, Latency and Context Window
  • Interacting Different LLMs with Chat UI: ChatGPT, LLama, Mixtral, Phi3
  • Installing and Running Llama and Gemma Models Using Ollama
  • Modernizing Enterprise Apps with AI-Powered LLM Capabilities
  • Designing the 'EShop Support App' with AI-Powered LLM Capabilities
  • Advanced Prompting Techniques: Zero-shot, One-shot, Few-shot, COT
  • Design Advanced Prompts for Ticket Detail Page in EShop Support App w/ Q&A Chat and RAG
  • The RAG Architecture: Ingestion with Embeddings and Vector Search
  • E2E Workflow of a Retrieval-Augmented Generation (RAG) - The RAG Workflow
  • End-to-End RAG Example for EShop Customer Support using OpenAI Playground
  • Fine-Tuning Methods: Full, Parameter-Efficient Fine-Tuning (PEFT), LoRA, Transfer
  • End-to-End Fine-Tuning a LLM for EShop Customer Support using OpenAI Playground
  • Choosing the Right Optimization – Prompt Engineering, RAG, and Fine-Tuning
  • Vector Database and Semantic Search with RAG
  • Explore Vector Embedding Models: OpenAI - text-embedding-3-small, Ollama - all-minilm
  • Explore Vector Databases: Pinecone, Chroma, Weaviate, Qdrant, Milvus, PgVector, Redis
  • Using LLMs and VectorDBs as Cloud-Native Backing Services in Microservices Architecture
  • Design EShop Support with LLMs, Vector Databases and Semantic Search
  • Design EShop Support with Azure Cloud AI Services: Azure OpenAI, Azure AI Search

In this course, you'll learn how to Design Generative AIArchitectures with integrating AI-Powered S/LLMs into EShop Support Enterprise Applications using Prompt Engineering, RAG, Fine-tuning and Vector DBs.

We will design Generative AI Architectures with below components;

  1. Small and Large Language Models (S/LLMs)

  2. Prompt Engineering

  3. Retrieval Augmented Generation (RAG)

  4. Fine-Tuning

  5. Vector Databases

We start with the basics and progressively dive deeper into each topic. We'll also follow LLM Augmentation Flow is a powerful framework that augments LLM results following the Prompt Engineering, RAG and Fine-Tuning.

Large Language Models (LLMs) module;

  • How Large Language Models (LLMs) works?

  • Capabilities of LLMs: Text Generation, Summarization, Q&A, Classification, Sentiment Analysis, Embedding Semantic Search, Code Generation

  • Generate Text with ChatGPT: Understand Capabilities and Limitations of LLMs (Hands-on)

  • Function Calling and Structured Output in Large Language Models (LLMs)

  • LLM Models: OpenAI ChatGPT, Meta Llama, Anthropic Claude, Google Gemini, Mistral Mixral, xAI Grok

  • SLM Models: OpenAI ChatGPT 4o mini, Meta Llama 3.2 mini, Google Gemma, Microsoft Phi 3.5

  • Interacting Different LLMs with Chat UI: ChatGPT, LLama, Mixtral, Phi3

  • Interacting OpenAI Chat Completions Endpoint with Coding

  • Installing and Running Llama and Gemma Models Using Ollama to run LLMs locally

  • Modernizing and Design EShop Support Enterprise Apps with AI-Powered LLM Capabilities

Prompt Engineering module;

  • Steps of Designing Effective Prompts: Iterate, Evaluate and Templatize

  • Advanced Prompting Techniques:Zero-shot, One-shot, Few-shot, Chain-of-Thought, Instruction and Role-based

  • Design Advanced Prompts for EShop Support – Classification, Sentiment Analysis, Summarization, Q&A Chat, and Response Text Generation

  • Design Advanced Prompts for Ticket Detail Page in EShop Support App w/ Q&A Chat and RAG

Retrieval-Augmented Generation (RAG) module;

  • The RAG Architecture Part 1: Ingestion with Embeddings and Vector Search

  • The RAG Architecture Part 2: Retrieval with Reranking and Context Query Prompts

  • The RAG Architecture Part 3: Generation with Generator and Output

  • E2E Workflow of a Retrieval-Augmented Generation (RAG) - The RAG Workflow

  • Design EShop Customer Support using RAG

  • End-to-End RAG Example for EShop Customer Support using OpenAI Playground

Fine-Tuning module;

  • Fine-Tuning Workflow

  • Fine-Tuning Methods: Full, Parameter-Efficient Fine-Tuning (PEFT), LoRA, Transfer

  • Design EShop Customer Support Using Fine-Tuning

  • End-to-End Fine-Tuning a LLM for EShop Customer Support using OpenAI Playground

Also, we will discuss

  • Choosing the Right Optimization – Prompt Engineering, RAG, and Fine-Tuning

Vector Database and Semantic Search with RAG module

  • What are Vectors, Vector Embeddings and Vector Database?

  • Explore Vector Embedding Models: OpenAI - text-embedding-3-small, Ollama - all-minilm

  • Semantic Meaning and Similarity Search: Cosine Similarity, Euclidean Distance

  • How Vector Databases Work: Vector Creation, Indexing, Search

  • Vector Search Algorithms: kNN, ANN, and Disk-ANN

  • Explore Vector Databases: Pinecone, Chroma, Weaviate, Qdrant, Milvus, PgVector, Redis

Lastly, we will Design EShopSupport Architecture with LLMs and Vector Databases

  • Using LLMs and VectorDBs as Cloud-Native Backing Services in Microservices Architecture

  • Design EShop Support with LLMs, Vector Databases and Semantic Search

  • Azure Cloud AI Services: Azure OpenAI, Azure AI Search

  • Design EShop Support with Azure Cloud AI Services: Azure OpenAI, Azure AI Search

This course is more than just learning Generative AI, it's a deep dive into the world of how to design Advanced AI solutions by integrating LLM architectures into Enterprise applications.

You'll get hands-on experience designing a complete EShop Customer Support application, including LLM capabilities like Summarization, Q&A, Classification, Sentiment Analysis, Embedding Semantic Search, Code Generation.

Taught by

Mehmet Ozkaya

Reviews

4.7 rating at Udemy based on 30 ratings

Start your review of Generative AI Architectures with LLM, Prompt, RAG, Vector DB

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.