Skip to main content
Vijil Evaluate is integrated with a number of leading LLM providers. To evaluate serverless LLM endpoints hosted on any of these, follow directions in the quickstart tutorial, but with different values for the model hub and model name.
ProviderModel HubOut-of-the-box model examplesDefault rate limit*
OpenAIopenaigpt-4.1, gpt-4.5-preview, gpt-4o, gpt-4o-mini, o1, o1-mini, o3-mini, gpt-4-turbo, gpt-4, gpt-3.5-turbo500 requests / 60s
Anthropicanthropicclaude-opus-4-0, claude-sonnet-4-0, claude-3-7-sonnet-latest, claude-3-5-sonnet-latest, claude-3-5-haiku-latest500 requests / 60s
Together AItogetherdeepseek-ai/DeepSeek-R1, meta-llama/Llama-3.3-70B-Instruct-Turbo, Qwen/QwQ-32B-Preview, google/gemma-2-27b-it, mistralai/Mixtral-8x22B-Instruct-v0.1600 requests / 60s
Mistral AImistralmistral-large-latest, mistral-saba-latest, ministral-3b-latest, ministral-8b-latest, mistral-small-latest, open-mixtral-8x22b300 requests / 60s
Fireworks AIfireworksllama-v3p2-1b-instruct, llama-v3p2-3b-instruct, llama-v3p2-11b-vision-instruct, llama-v3p2-90b-vision-instruct, mixtral-8x22b-instruct, qwen2p5-72b-instruct, deepseek-r1600 requests / 60s
NVIDIA NIMnvidianvidia/llama3-chatqa-1.5-8b, nvidia/llama3-chatqa-1.5-70b, nvidia/nemotron-4-340b-instruct60 requests / 60s
Google Cloud Vertex AIvertexgoogle/gemini-2.5-pro, google/gemini-2.0-flash-001, google/gemini-1.5-flash-001, google/gemini-1.5-pro-001, google/gemini-1.0-pro-00260 requests / 60s
AWS Bedrockbedrockus.amazon.nova-lite-v1:0, us.amazon.nova-micro-v1:0, us.amazon.nova-pro-v1:0, us.anthropic.claude-3-7-sonnet-20250219-v1:0, anthropic.claude-3-5-sonnet-20241022-v2:0, meta.llama3-1-70b-instruct-v1:0, meta.llama3-1-405b-instruct-v1:0, mistral.mistral-large-2407-v1:060 requests / 60s
AWS Bedrock AgentsbedrockAgentsBedrock-hosted agents configured by agent_id / agent_alias_id30 requests / 60s
Google Cloud (Agentforce)agentforceAgentforce-hosted agents (via Agentforce configuration)30 requests / 60s
AzureazureAzure-hosted deployments (models configured in your Azure account)60 requests / 60s
DigitalOceandigitaloceanDigitalOcean-hosted deployments (models configured in your DO account)30 requests / 60s
OpenRouteropenroutermicrosoft/phi-4, google/gemini-2.5-flash-preview, x-ai/grok-3-beta60 requests / 60s
Groqgroqllama-3.1-8b-instant, llama-3.3-70b-versatile, openai/gpt-oss-120b, openai/gpt-oss-20b30 requests / 60s
Vijil also supports a number of other cloud services, giving you the flexibility of evaluating agents accessible through custom endpoints. *Default rate limits are the initial Vijil-side quotas (in requests per 60 seconds) and can be further constrained by your provider account and perโ€“API key configuration in Vijil.

Google Vertex AI

Learn more about integrating Google Vertex AI

DigitalOcean

Learn more about integrating DigitalOcean

AWS Bedrock

Learn more about integrating AWS Bedrock
Last modified on March 19, 2026