Table of contents
Official Content
  • This documentation is valid for:

List of supported models via /chat Api

Module Provider Model Full Name Chat Support Function Calling support Environment support
saia.models.openai OpenAI
  • openai/gpt-4o
  • openai/gpt-4o-mini
  • openai/gpt-4o-2024-11-20
  • openai/gpt-4.1
  • openai/gpt-4.1-mini
  • openai/gpt-4.1-nano
  • openai/o1(1)
  • openai/o1-pro(1)
  • openai/o3(1)
  • openai/o3-pro(1)
  • openai/o3-mini(1)
  • openai/o4-mini(1)
  • openai/codex-mini-latest
  • openai/o4-mini-deep-research(3)
  • openai/o3-deep-research(3)
  • Beta
  • Production
  • openai/chatgpt-4o-latest
  • openai/gpt-4o-search-preview(2)
  • openai/gpt-4o-mini-search-preview(2)
 
saia.models.googlevertexai Google VertexAI
  • vertex_ai/gemini-2.5-pro
  • vertex_ai/gemini-2.5-flash
  • vertex_ai/gemini-2.0-flash
  • vertex_ai/gemini-2.0-flash-lite
  • vertex_ai/gemini-2.0-flash-001
  • vertex_ai/claude-opus-4-20250514
  • vertex_ai/claude-sonnet-4-20250514
  • vertex_ai/claude-3-7-sonnet-20250219
  • vertex_ai/claude-3-5-sonnet-20240620
  • vertex_ai/claude-3-5-sonnet-v2-20241022
  • vertex_ai/claude-3-5-haiku-20241022
  • vertex_ai/claude-3-opus-20240229(5)
  • vertex_ai/claude-3-haiku-20240307(5)
  • vertex_ai/meta-llama-4-maverick-17b-128e-instruct-maas
  • vertex_ai/meta-llama-4-scout-17b-16e-instruct-maas
  • vertex_ai/mistral-small-2503
  • vertex_ai/mistral-large-2411
  • Beta
  • Production
  • vertex_ai/codestral-2501
 
saia.models.azure Azure OpenAI
  • azure/gpt-4.1
  • azure/gpt-4.1-mini
  • azure/gpt-4.1-nano
  • azure/gpt-4o
  • azure/gpt-4o-mini
  • azure/o1
  • azure/o3-mini
  • azure/o4-mini
  • Beta
  • Production
  • azure/o1-mini
 
saia.models.anthropic Anthropic
  • anthropic/claude-3-opus-20240229(5)
  • anthropic/claude-3-haiku-20240307(5)
  • anthropic/claude-3-5-sonnet-20240620
  • anthropic/claude-3-5-sonnet-20241022
  • anthropic/claude-3-5-haiku-20241022
  • anthropic/claude-3-7-sonnet-latest
  • anthropic/claude-opus-4-20250514
  • anthropic/claude-sonnet-4-20250514
  • Beta
  • Production
saia.models.awsbedrock AWS Bedrock
  • awsbedrock/us.anthropic.claude-opus-4-20250514-v1:0
    awsbedrock/us.anthropic.claude-sonnet-4-20250514-v1:0
  • awsbedrock/anthropic.claude-3-7-sonnet
  • awsbedrock/anthropic.claude-3.5-sonnet-v2
  • awsbedrock/anthropic.claude-3.5-sonnet
  • awsbedrock/anthropic.claude-3.5-haiku
  • awsbedrock/anthropic.claude-3-haiku(5)
  • awsbedrock/anthropic.claude-3-opus(5)
  • awsbedrock/meta.llama3-8b
  • awsbedrock/meta.llama3-70b
  • awsbedrock/amazon.titan-lite-v1
  • awsbedrock/amazon.titan-express-v1
  • awsbedrock/cohere.command
  • awsbedrock/meta.llama3-1-70b
  • awsbedrock/meta.llama3-1-405b
  • awsbedrock/amazon.nova-pro-v1:0
  • awsbedrock/amazon.nova-lite-v1:0
  • awsbedrock/amazon.nova-micro-v1:0
  • awsbedrock/meta.llama3-2-1b
  • awsbedrock/meta.llama3-2-3b
  • awsbedrock/meta.llama3-2-11b
  • awsbedrock/meta.llama3-2-90b
  • Beta
  • Production
  • awsbedrock/us.deepseek.r1-v1:0
 
saia.models.xai xAI
  • xai/grok-4
  • xai/grok-3
  • xai/grok-3-fast
  • xai/grok-3-mini
  • xai/grok-3-mini-fast
  • xai/grok-2-vision-1212
  • Beta
  • Production
saia.models.cohere Cohere
  • cohere/command-r
  • cohere/command-r-08-2024
  • cohere/command-r-plus
  • cohere/command-r-plus-08-2024
  • cohere/command-r7b-12-2024
  • cohere/command-a-03-2025
  • Beta
  • Production
saia.models.azure.foundry Azure AI Foundry
  • azure_ai_foundry/grok-3
  • azure_ai_foundry/grok-3-mini
  • azure_ai_foundry/gpt-4.1
  • azure_ai_foundry/gpt-4.1-mini
  • azure_ai_foundry/DeepSeek-V3-0324
  • azure_ai_foundry/DeepSeek-R1-0528
  • Beta
  • Production
  • azure_ai_foundry/DeepSeek-R1
  • azure_ai_foundry/Phi-4
  • azure_ai_foundry/Phi-4-mini-instruct
  • azure_ai_foundry/Phi-4-mini-reasoning
  • azure_ai_foundry/Phi-4-multimodal-instruct
 

saia.models.openrouter

OpenRouter

  • openrouter/auto
  • openrouter/qwen3-32b
  • openrouter/qwen3-235b-a22b
  • openrouter/deepseek-chat-v3-0324

  • Beta
  • openrouter/qwen3-8b:free
  • openrouter/qwen3-14b:free
  • openrouter/qwen3-30b-a3b:free
  • openrouter/qwen3-32b:free
  • openrouter/qwen3-235b-a22b:free
 
saia.models.mistral Mistral AI
  • mistral/magistral-medium-latest
  • mistral/mistral-medium-latest
  • mistral/codestral-latest
  • mistral/mistral-saba-latest
  • mistral/mistral-large-latest
  • mistral/pixtral-large-latest
  • mistral/ministral-3b-latest
  • mistral/ministral-8b-latest
  • mistral/devstral-small-latest
  • mistral/mistral-small-latest
  • mistral/pixtral-12b-2409
  • mistral/open-mistral-nemo
  • Beta
  • mistral/magistral-small-latest
 
saia.models.deepseek DeepSeek
  • deepseek/deepseek-chat
  • deepseek/deepseek-reasoner
  • Beta
saia.models.groq Groq
  • groq/moonshotai-kimi-k2-instruct
  • groq/llama-3.3-70b-versatile
  • groq/llama-3.1-8b-instant
  • groq/meta-llama-4-scout-17b-16e-instruct
  • groq/meta-llama-4-maverick-17b-128e-instruct
  • groq/qwen-qwq-32b
  • groq/qwen3-32b
  • groq/deepseek-r1-distill-llama-70b
  • groq/mistral-saba-24b
  • Beta
saia.models.nvidia NVidia
  • nvidia/nvidia.nemotron-mini-4b-instruct
  • nvidia/meta.llama-3.1-8b-instruct
  • nvidia/meta.llama-3.1-70b-instruct
  • nvidia/meta.llama-3.1-405b-instruct
  • nvidia/meta.llama-3.2-3b-instruct
  • nvidia/meta-llama-4-scout-17b-16e-instruct
  • nvidia/llama-3.3-nemotron-super-49b-v1
  • nvidia/llama-3.1-nemotron-70b-instruct
  • Beta
  • nvidia/meta.llama-3.2-1b-instruct
  • nvidia/llama-3.1-nemotron-ultra-253b-v1
  • nvidia/meta-llama-4-maverick-17b-128e-instruct
  • nvidia/deepseek-ai-deepseek-r1
 
saia.models.sambanova SambaNova
  • sambanova/Meta-Llama-3.3-70B-Instruct
  • sambanova/Llama-4-Maverick-17B-128E-Instruct
  • Beta
  • sambanova/DeepSeek-R1-Distill-Llama-70B
 
saia.models.cerebras Cerebras
  • cerebras/llama3.1-8b
  • cerebras/llama-3.3-70b
  • Beta
  • cerebras/llama-4-scout-17b-16e-instruct
 
saia.models.inception Inception Labs
  • inception/mercury(4)
  • inception/mercury-coder(4)
 
  • Beta

(1) - To use these models the temperature must be set to 1, check Reasoning models.
(2) - These models do not support the Temperature parameter in the request body.
(3) - This model is only available via Responses API.
(4) - difussion LLM (dLLM).
(5) - This model will be deprecated soon. Please refer to Deprecated Models for more information.

Globant Enterprise AI LLM consumption limits in SaaS mode

When using Globant Enterprise AI in SaaS mode, you have a monthly limit of 11,000 requests for the following LLMs:

  • OpenAI:
    • GPT-4.1
    • GPT-4o
    • o1
    • o1-pro
    • o3-pro
    • o3-mini
    • o4-mini
  • Google Vertex AI:
    • Gemini Pro 2.5
    • Gemini Pro 1.5
    • Claude 4 Opus
    • Calude 4 Sonnet
    • Claude 3.7 Sonnet
  • AWS Bedrock:
    • Claude 4 Opus
    • Calude 4 Sonnet
    • Claude 3.7 Sonnet
    • Claude 3.5 Sonnet (v1 and v2)
    • Llama 3.1 Instruct (405b)

For any other LLMs or models, pricing and usage limits will be evaluated on a case-by-case basis.

See Also

Last update: June 2025 | © GeneXus. All rights reserved. GeneXus Powered by Globant