📄️ OpenAI
LiteLLM supports OpenAI Chat + Text completion and embedding calls.
📄️ Azure OpenAI
API Keys, Params
📄️ Huggingface
LiteLLM supports the following types of Huggingface models:
📄️ Ollama
LiteLLM supports all models from Ollama
📄️ VertexAI - Google
Pre-requisites
📄️ PaLM API - Google
Pre-requisites
📄️ Anthropic
LiteLLM supports
📄️ AWS Sagemaker
LiteLLM supports Llama2 on Sagemaker
📄️ AWS Bedrock
Anthropic, Amazon Titan, A121 LLMs are Supported on Bedrock
📄️ Anyscale
https://app.endpoints.anyscale.com/
📄️ Perplexity AI (pplx-api)
https://www.perplexity.ai
📄️ VLLM
LiteLLM supports all models on VLLM.
📄️ DeepInfra
https://deepinfra.com/
📄️ AI21
LiteLLM supports j2-light, j2-mid and j2-ultra from AI21.
📄️ NLP Cloud
LiteLLM supports all LLMs on NLP Cloud.
📄️ Replicate
LiteLLM supports all models on Replicate
📄️ Cohere
API KEYS
📄️ Together AI
LiteLLM supports all models on Together AI.
📄️ Aleph Alpha
LiteLLM supports all models from Aleph Alpha.
📄️ Baseten
LiteLLM supports any Text-Gen-Interface models on Baseten.
📄️ OpenRouter
LiteLLM supports all the text models from OpenRouter
📄️ Custom API Server (OpenAI Format)
LiteLLM allows you to call your custom endpoint in the OpenAI ChatCompletion format
📄️ Petals
Petals//github.com/bigscience-workshop/petals