Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

Linux Foundation

Boosting LLM Development and Training Efficiency: Automated Parallelization with MindSpore

Linux Foundation via YouTube

Overview

Explore automated parallelization techniques for boosting large language model (LLM) development and training efficiency using MindSpore in this 44-minute conference talk by Yiren Xing from Huawei. Learn about an innovative approach that allows developers to focus on algorithm research without the need for intrusive model modifications. Discover how distributed training on large-scale clusters can be achieved through simple strategy configurations. Gain insights into MindSpore's hyperparameter search model for automatically finding optimal parallelization strategies, which can achieve 90%-110% of expert tuning performance. Understand how this method significantly reduces model modification time while efficiently accelerating LLM training. The presentation covers both the challenges of large-scale distributed parallel training and the solutions offered by automated parallelization, making it valuable for AI researchers and developers working with large language models.

Syllabus

Boosting LLM Development & Training Efficiency: Automated Parallelization with MindSpore- Yiren Xing

Taught by

Linux Foundation

Reviews

Start your review of Boosting LLM Development and Training Efficiency: Automated Parallelization with MindSpore

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.