📄️ Huggingface
LiteLLM supports Huggingface Inference Endpoints that uses the text-generation-inference format.
📄️ OpenAI
LiteLLM supports OpenAI Chat + Text completion and embedding calls.
📄️ Azure
LiteLLM supports Azure Chat + Embedding calls.
📄️ Google Palm/VertexAI
LiteLLM supports chat-bison, chat-bison@001, text-bison, text-bison@001
📄️ Anthropic
LiteLLM supports Claude-1, 1.2 and Claude-2.
📄️ VLLM
LiteLLM supports all models on VLLM.
📄️ AI21
LiteLLM supports j2-light, j2-mid and j2-ultra from AI21.
📄️ Replicate
LiteLLM supports all models on Replicate
📄️ Cohere
LiteLLM supports 'command', 'command-light', 'command-medium', 'command-medium-beta', 'command-xlarge-beta', 'command-nightly' models from Cohere.
📄️ Together AI
LiteLLM supports all models on Together AI.
📄️ AWS Sagemaker
LiteLLM supports Llama2 on Sagemaker
📄️ Aleph Alpha
LiteLLM supports all models from Aleph Alpha.
📄️ Baseten
LiteLLM supports any Text-Gen-Interface models on Baseten.
📄️ OpenRouter
LiteLLM supports all the text models from OpenRouter
📄️ Ollama
LiteLLM supports all models from Ollama
📄️ Custom LLM API-Endpoints
LiteLLM supports Custom deploy api endpoints
📄️ OpenAI Proxy Servers (ChatCompletion)
LiteLLM allows you to call your OpenAI ChatCompletion proxy server