Hymalaia can be configured to use a custom model server through REST requests. This guide explains how to set up and integrate your own model server with Hymalaia.
Hymalaia supports making requests to arbitrary model servers via REST API endpoints. You can optionally include an access token for authentication. For custom request formats or response handling, you may need to update and rebuild the Hymalaia containers.
To make Hymalaia compatible with your custom model server, you’ll need to implement a minimal interface that can support any arbitrary LLM Model Server. The process involves:
The default implementation provides a reference that you can modify according to your needs.
As a practical example, you can set up Hymalaia with a self-hosted Llama-2-13B-chat-GGML model using a custom FastAPI server.
You can try this implementation using Google Colab for GPU access. However, please note that Colab is not recommended for production deployments.
For detailed implementation steps and code examples, refer to our Medium blog post.
Common issues and solutions:
For additional support or questions, please refer to our documentation or community forums.
Hymalaia can be configured to use a custom model server through REST requests. This guide explains how to set up and integrate your own model server with Hymalaia.
Hymalaia supports making requests to arbitrary model servers via REST API endpoints. You can optionally include an access token for authentication. For custom request formats or response handling, you may need to update and rebuild the Hymalaia containers.
To make Hymalaia compatible with your custom model server, you’ll need to implement a minimal interface that can support any arbitrary LLM Model Server. The process involves:
The default implementation provides a reference that you can modify according to your needs.
As a practical example, you can set up Hymalaia with a self-hosted Llama-2-13B-chat-GGML model using a custom FastAPI server.
You can try this implementation using Google Colab for GPU access. However, please note that Colab is not recommended for production deployments.
For detailed implementation steps and code examples, refer to our Medium blog post.
Common issues and solutions:
For additional support or questions, please refer to our documentation or community forums.