Explore hardware requirements, software configurations, performance metrics, and cost considerations for running Llama-2 70b and Mixtral 8x7b unquantized inference with low concurrency in this informative 52-minute conference talk. Analyze benchmark data from GitHub to compare different frameworks and hardware setups, including on-premises and cloud-based solutions in the USA. Gain insights into the speed and cost advantages of open-source software (OSS) LLMs compared to closed APIs like OpenAI. Learn how to configure servers and replicate benchmarks through practical code examples. Acquire valuable knowledge to jumpstart your journey in implementing high-quality OSS LLM inference for small organizations in 2024, focusing on optimal hardware and software choices for efficient performance.
Overview
Syllabus
HW, SW, Performance and Costs for Llama-2 70b and Mixtral 8x7b LLM Inference with Low...- Ivan Baldo
Taught by
Linux Foundation