Skip to content

Deploy LLM Inference APIs anywhere within seconds

Deploy any open source Large Language model (LLM) from Huggingface or Ollama Registry to Akash network with zero configuration

Streamling LLM Infrence API Deployment on any cloud plaform via Ollama, vLLM and Llama-cpp

Deploy open source models within seconds on Akash Network:

  1. Select the Model of your choice from Huggingface or Ollama registry.
  2. Hit Deploy to get a properly configured compose file with infernce api of the selected model.
  3. Copy the compose and use it to deploy on GCP, Azure or AWS.

The deployment process is fully automated and managed and needs no configuration, which ensures that the service is highly available, scalable, and secure.


Endpoint Documentation

Click here to see the user docs