Table of contents
Official Content
  • This documentation is valid for:

List of supported models via /chat Api

Module Provider Model Full Name Chat Support Function Calling support Environment support
saia.models.openai OpenAI
  • openai/gpt-4o
  • openai/gpt-4o-mini
  • openai/gpt-4o-2024-11-20
  • openai/gpt-4.5-preview(4)
  • openai/gpt-4.1
  • openai/gpt-4.1-mini
  • openai/gpt-4.1-nano
  • openai/o1(1)
  • openai/o1-pro(1)
  • openai/o3(1)
  • openai/o3-pro(1)
  • openai/o3-mini(1)
  • openai/o4-mini(1)
  • openai/codex-mini-latest
  • openai/o4-mini-deep-research(3)
  • openai/o3-deep-research(3)

openai/* (any openai model)

  • Beta
  • Production
  • openai/chatgpt-4o-latest
  • openai/gpt-4o-search-preview(2)
  • openai/gpt-4o-mini-search-preview(2)
 
saia.models.googlevertexai Google VertexAI
  • vertex_ai/gemini-2.5-pro
  • vertex_ai/gemini-2.5-flash
  • vertex_ai/gemini-2.5-pro-preview-05-06(4)
  • vertex_ai/gemini-2.5-flash-preview-04-17(4)
  • vertex_ai/gemini-2.0-flash
  • vertex_ai/gemini-2.0-flash-lite
  • vertex_ai/gemini-2.0-flash-001
  • vertex_ai/claude-opus-4-20250514
  • vertex_ai/claude-sonnet-4-20250514
  • vertex_ai/claude-3-7-sonnet-20250219
  • vertex_ai/claude-3-5-sonnet-20240620
  • vertex_ai/claude-3-opus-20240229
  • vertex_ai/claude-3-haiku-20240307
  • vertex_ai/claude-3-5-sonnet-v2-20241022
  • vertex_ai/claude-3-5-haiku-20241022
  • vertex_ai/meta-llama-4-maverick-17b-128e-instruct-maas
  • vertex_ai/meta-llama-4-scout-17b-16e-instruct-maas
  • vertex_ai/mistral-small-2503
  • vertex_ai/mistral-large-2411
  • Beta
  • Production
  • vertex_ai/codestral-2501
 
saia.models.azure Azure OpenAI
  • azure/gpt-4.1
  • azure/gpt-4.1-mini
  • azure/gpt-4.1-nano
  • azure/gpt-4o
  • azure/gpt-4o-mini
  • azure/o1
  • azure/o3-mini
  • azure/o4-mini
  • Beta
  • Production
  • azure/o1-mini
 
saia.models.anthropic Anthropic
  • anthropic/claude-3-opus-20240229
  • anthropic/claude-3-haiku-20240307
  • anthropic/claude-3-5-sonnet-20240620
  • anthropic/claude-3-5-sonnet-20241022
  • anthropic/claude-3-5-haiku-20241022
  • anthropic/claude-3-7-sonnet-latest
  • anthropic/claude-opus-4-20250514
  • anthropic/claude-sonnet-4-20250514
  • Beta
  • Production
saia.models.awsbedrock AWS Bedrock
  • awsbedrock/us.anthropic.claude-opus-4-20250514-v1:0
    awsbedrock/us.anthropic.claude-sonnet-4-20250514-v1:0
  • awsbedrock/anthropic.claude-3-7-sonnet
  • awsbedrock/anthropic.claude-3-haiku
  • awsbedrock/anthropic.claude-3.5-sonnet
  • awsbedrock/anthropic.claude-3-opus
  • awsbedrock/meta.llama3-8b
  • awsbedrock/meta.llama3-70b
  • awsbedrock/amazon.titan-lite-v1
  • awsbedrock/amazon.titan-express-v1
  • awsbedrock/cohere.command
  • awsbedrock/meta.llama3-1-70b
  • awsbedrock/meta.llama3-1-405b
  • awsbedrock/anthropic.claude-3.5-sonnet-v2
  • awsbedrock/anthropic.claude-3.5-haiku
  • awsbedrock/amazon.nova-pro-v1:0
  • awsbedrock/amazon.nova-lite-v1:0
  • awsbedrock/amazon.nova-micro-v1:0
  • awsbedrock/meta.llama3-2-1b
  • awsbedrock/meta.llama3-2-3b
  • awsbedrock/meta.llama3-2-11b
  • awsbedrock/meta.llama3-2-90b
  • Beta
  • Production
  • awsbedrock/us.deepseek.r1-v1:0
 
saia.models.xai xAI
  • xai/grok-3
  • xai/grok-3-fast
  • xai/grok-3-mini
  • xai/grok-3-mini-fast
  • xai/grok-2-vision-1212
  • Beta
  • Production
saia.models.cohere Cohere
  • cohere/command-r
  • cohere/command-r-08-2024
  • cohere/command-r-plus
  • cohere/command-r-plus-08-2024
  • cohere/command-r7b-12-2024
  • cohere/command-a-03-2025(5)
  • Beta
  • Production
saia.models.azure.foundry Azure AI Foundry
  • azure_ai_foundry/grok-3
  • azure_ai_foundry/grok-3-mini
  • azure_ai_foundry/gpt-4.1
  • azure_ai_foundry/gpt-4.1-mini
  • azure_ai_foundry/DeepSeek-V3-0324
  • azure_ai_foundry/DeepSeek-R1-0528
  • Beta
  • Production
  • azure_ai_foundry/DeepSeek-R1
  • azure_ai_foundry/Phi-4
  • azure_ai_foundry/Phi-4-mini-instruct
  • azure_ai_foundry/Phi-4-mini-reasoning
  • azure_ai_foundry/Phi-4-multimodal-instruct
 

saia.models.openrouter

OpenRouter

  • openrouter/auto
  • openrouter/qwen3-32b
  • openrouter/qwen3-235b-a22b
  • openrouter/deepseek-chat-v3-0324

  • Beta
  • openrouter/qwen3-8b:free
  • openrouter/qwen3-14b:free
  • openrouter/qwen3-30b-a3b:free
  • openrouter/qwen3-32b:free
  • openrouter/qwen3-235b-a22b:free
 
saia.models.mistral Mistral AI
  • mistral/magistral-medium-latest
  • mistral/mistral-medium-latest
  • mistral/codestral-latest
  • mistral/mistral-saba-latest
  • mistral/mistral-large-latest
  • mistral/pixtral-large-latest
  • mistral/ministral-3b-latest
  • mistral/ministral-8b-latest
  • mistral/devstral-small-latest
  • mistral/mistral-small-latest
  • mistral/pixtral-12b-2409
  • mistral/open-mistral-nemo
  • Beta
  • mistral/magistral-small-latest
 
saia.models.deepseek DeepSeek
  • deepseek/deepseek-chat
  • deepseek/deepseek-reasoner
  • Beta
saia.models.groq Groq
  • groq/llama-3.3-70b-versatile
  • groq/llama-3.1-8b-instant
  • groq/meta-llama-4-scout-17b-16e-instruct
  • groq/meta-llama-4-maverick-17b-128e-instruct
  • groq/qwen-qwq-32b
  • groq/qwen3-32b
  • groq/deepseek-r1-distill-llama-70b
  • groq/mistral-saba-24b
  • Beta
saia.models.nvidia NVidia
  • nvidia/nvidia.nemotron-mini-4b-instruct
  • nvidia/meta.llama-3.1-8b-instruct
  • nvidia/meta.llama-3.1-70b-instruct
  • nvidia/meta.llama-3.1-405b-instruct
  • nvidia/meta.llama-3.2-3b-instruct
  • nvidia/meta-llama-4-scout-17b-16e-instruct
  • nvidia/llama-3.3-nemotron-super-49b-v1
  • nvidia/llama-3.1-nemotron-70b-instruct
  • Beta
  • nvidia/meta.llama-3.2-1b-instruct
  • nvidia/llama-3.1-nemotron-ultra-253b-v1
  • nvidia/meta-llama-4-maverick-17b-128e-instruct
  • nvidia/deepseek-ai-deepseek-r1
 
saia.models.sambanova SambaNova
  • sambanova/Meta-Llama-3.3-70B-Instruct
  • sambanova/Llama-4-Maverick-17B-128E-Instruct
  • Beta
  • sambanova/DeepSeek-R1-Distill-Llama-70B
 
saia.models.cerebras Cerebras
  • cerebras/llama3.1-8b
  • cerebras/llama-3.3-70b
  • Beta
  • cerebras/llama-4-scout-17b-16e-instruct
 
saia.models.inception Inception Labs
  • inception/mercury(6)
  • inception/mercury-coder(6)
 
  • Beta

(1) - To use these models the temperature must be set to 1, check Reasoning models.
(2) - These models do not support the Temperature parameter in the request body.
(3) - This model is only available via Responses API.
(4) - This model will be deprecated soon. Please refer to Deprecated Models for further information.
(5) - Available since June 2025 Release.
(6) - difussion LLM (dLLM)

Globant Enterprise AI LLM consumption limits in SaaS mode

When using Globant Enterprise AI in SaaS mode, you have a monthly limit of 11,000 requests for the following LLMs:

  • OpenAI:
    • GPT-4.5
    • GPT-4.1
    • GPT-4o
    • o1
    • o1-pro
    • o3-pro
    • o3-mini
    • o4-mini
  • Google Vertex AI:
    • Gemini Pro 2.5
    • Gemini Pro 1.5
    • Claude 4 Opus
    • Calude 4 Sonnet
    • Claude 3.7 Sonnet
  • AWS Bedrock:
    • Claude 4 Opus
    • Calude 4 Sonnet
    • Claude 3.7 Sonnet
    • Claude 3.5 Sonnet (v1 and v2)
    • Llama 3.1 Instruct (405b)

For any other LLMs or models, pricing and usage limits will be evaluated on a case-by-case basis.

See Also

Last update: June 2025 | © GeneXus. All rights reserved. GeneXus Powered by Globant