Configure Hymalaia to use HuggingFace Inference APIs
To use HuggingFace Inference APIs with Hymalaia, follow the instructions below.
You must have a Pro Account with HuggingFace to obtain an API key.
⚠️ Note: As of November 2023, HuggingFace no longer supports very large models (over 10GB) like
LLaMA-2-70B
on the Pro Plan. You’ll need to:
- Use a dedicated Inference Endpoint (paid)
- Or subscribe to an Enterprise Plan
The Pro Plan still works with smaller models, but these may yield suboptimal results for Hymalaia.
HFAccessToken
).Refer to your deployment-specific documentation for setting environment variables.
To configure Hymalaia for next-token generation using HuggingFace’s Inference API:
These custom providers allow Hymalaia to route prompt completion requests to the HuggingFace-hosted model endpoint.
For more detailed setup and environment configuration examples, refer to the Model Configs.
Configure Hymalaia to use HuggingFace Inference APIs
To use HuggingFace Inference APIs with Hymalaia, follow the instructions below.
You must have a Pro Account with HuggingFace to obtain an API key.
⚠️ Note: As of November 2023, HuggingFace no longer supports very large models (over 10GB) like
LLaMA-2-70B
on the Pro Plan. You’ll need to:
- Use a dedicated Inference Endpoint (paid)
- Or subscribe to an Enterprise Plan
The Pro Plan still works with smaller models, but these may yield suboptimal results for Hymalaia.
HFAccessToken
).Refer to your deployment-specific documentation for setting environment variables.
To configure Hymalaia for next-token generation using HuggingFace’s Inference API:
These custom providers allow Hymalaia to route prompt completion requests to the HuggingFace-hosted model endpoint.
For more detailed setup and environment configuration examples, refer to the Model Configs.