Skip to content
Menu
Home
Get Started
Get started with Intel® Gaudi® AI accelerators
Using Hugging Face
Tutorials
Get Access
Get Optimized
Model Performance Data
Generative AI and Large Language Models
Software Setup and Installation
Kernel Libraries
Intel® Gaudi® accelerators on Premise
Videos
Documentation
Installation Guide
PyTorch User Guide
DeepSpeed User Guide
Catalog
Containers
Generative
Large Language Models
Computer Vision
Natural Language Processing
PyTorch
Forum
Explore More
Intel Gaudi Blog
Events and Webinars
FAQs
Support
Menu
Home
Get Started
Get started with Intel® Gaudi® AI accelerators
Using Hugging Face
Tutorials
Get Access
Get Optimized
Model Performance Data
Generative AI and Large Language Models
Software Setup and Installation
Kernel Libraries
Intel® Gaudi® accelerators on Premise
Videos
Documentation
Installation Guide
PyTorch User Guide
DeepSpeed User Guide
Catalog
Containers
Generative
Large Language Models
Computer Vision
Natural Language Processing
PyTorch
Forum
Explore More
Intel Gaudi Blog
Events and Webinars
FAQs
Support
Home
»
Inference
Intel® Gaudi® AI Accelerators Blog
/ Inference
08/16/2023
Optimizing Large Language Model Inference on Gaudi2 with Hugging Face Optimum-Habana
We have optimized additional Large Language Models on Hugging Face using the Optimum Habana library.
Read more
DeepSpeed
,
Hugging Face
,
Inference
02/14/2023
BLOOM 176B Inference on Habana Gaudi2
With Habana’s SynapseAI 1.8.0 release support of DeepSpeed Inference, users can run inference on large language models, including BLOOM 176B.
Read more
BLOOM
,
DeepSpeed
,
Inference