Official Content
  • This documentation is valid for:

Supported Chat Models

 

(BETA Environment)

Module Provider Proxy API Model Chat API Assistant Module RAG Module Frontend
saia.models.openai OpenAI * (1)
saia.models.googlegenai Google Generative AI gemini-1.0-pro (2)
saia.models.googlevertexai Google VertexAI gemini-1.0-pro (3)

gemini-1.5-pro-preview-0409

pending Anthropic claude-3-opus-20240229 (4)

claude-3-sonnet-20240229

               
pending Azure OpenAI * (1) (5)
saia.models.nvidia-nemo NVidia Nemo NVidia Models (2)        

(1) - Needs to be enabled by request.
(2) - Set the RAG LLM Provider as google.
(3) - Set the RAG LLM Provider as googlevertexai.
(4) - Set the RAG LLM Provider as anthropic.
(5) - Set the RAG LLM Provider as azureopenai.

NVidia Models Support (2)

Model Proxy API Chat API Assistant Module RAG Module(2) Frontend
code-llama-70B
code-llama-13B
code-llama-34B
NV-Llama2-70B-RLHF
NV-Llama2-70B-SteerLM-Chat
mamba-chat
Nemotron-3-8B-QA (1)
Nemotron-3-8B-Chat-SteerLM
Mixtral-8x7B-Instruct
gemma-2B
gemma-7B
Yi-34B
NeVA-22B (1)
Fuyu-8B (1)
Llama-2-70B
Llama-2-13B
meta/llama3-8b-instruct (4)   (3) (3) (3)
meta/llama3-70b-instruct  
snowflake/arctic   (3)
databricks/dbrx-instruct   (3)
mistralai/mistral-7b-instruct-v0.2   (3)
mistralai/mistral-large  
mistralai/mixtral-8x22b-instruct-v0.1  
microsoft/phi-3-mini-128k-instruct   (3)

(1) - Not a chat model
(2) - The provider to set for the RAG LLM Settings section must be: nvidia.nemo and the Model Name must match one of the supported models detailed.
(3) - Available since Nvidia module Revision #4.
(4) - Evaluation pending

 

Last update: November 2024 | © GeneXus. All rights reserved. GeneXus Powered by Globant