Supported LLMs
Ax supports all the top LLM providers and models, along with their advanced capabilities, such as function calling, multi-modal, streaming, and JSON.
Our defaults, including default models, are selected to ensure solid agent performance.
OpenAI
Azure OpenAI
Azure requires you to set a resource name and a deployment name
https://learn.microsoft.com/en-us/azure/ai-services/openai/how-to/create-resource?pivots=web-portal
Together
Together runs a diverse array of open-source models, each designed for a specific use case. This variety ensures that you can find the perfect model for your needs.
https://docs.together.ai/docs/inference-models
Anthropic
Groq
Groq uses specialized hardware to serve open-source models with the lowest latency. It supports a small number of good models.
Google Gemini
An excellent model family with very long context lengths at the lowest price points. Gemini has built-in support for compute (code execution); their models can write and run code in the backend if needed.
Cohere
Huggingface
Mistral
Deepseek
Deepseek is an LLM provider from China that has excellent models.
Ollama
Ollama is an engine for running open-source models locally on your laptop. We default to nous-hermes21
for inference and all-minilm
for embedding.