Hugging Face가 Intel Gaudi 하드웨어를 Text Generation Inference에 네이티브 통합으로 별도 포크 제거 및 최신 기능 동시 지원
🚀 Accelerating LLM Inference with TGI on Intel Gaudi
🚀 Accelerating LLM Inference with TGI on Intel Gaudi
Introducing multi-backends (TRT-LLM, vLLM) support for Text Generation Inference
Benchmarking Text Generation Inference
Hugging Face on AMD Instinct MI300 GPU
Ryght’s Journey to Empower Healthcare and Life Sciences with Expert Support from Hugging Face
From OpenAI to Open LLMs with Messages API on Hugging Face
Hugging Face Text Generation Inference available for AWS Inferentia2
Inference for PROs
Introducing the Hugging Face LLM Inference Container for Amazon SageMaker