Home » DeepSpeed

Habana Developer Blog

/ DeepSpeed
In this tutorial, we will demonstrate fine tuning a GPT2 model on Habana Gaudi AI processors using Hugging Face optimum-habana library with DeepSpeed. What is fine tuning? Training models from scratch can be expensive, especially ... Read more
One of the key challenges in Large Language Model (LLM) training is reducing the memory requirements needed for training without sacrificing compute/communication efficiency and model accuracy.  DeepSpeed [2] is a popular deep learning software library ... Read more
Sign up for the latest Habana developer news, events, training, and updates.