INTEGRATED WITH ALL LEADING PROVIDERS

Multi-model support

Vertesia provides out-of-the-box connectivity for hundreds of generative AI models across all of the leading inference providers, and with our universal prompting, you can write one set of prompts which is instantly usable across any supported model
OpenAI-600x240
Gemini-600x240
Amazon-Nova-600x240
Google-Gemma-600x240
Amazon-Titan-600x240
Claude-600x240
IBM-Granite-600x240
Cohere-600x240
Meta-600x240
AI21_Labs_600x240
Mistral-600x240
DeepSeek-600x240

We believe that multi-model support is foundational to your success with generative AI (GenAI)

WHY THIS MATTERS

In our experience, there is a right model for every generative AI (GenAI) task and activity

For optimal results with complex content, workflows, or agents, you might need to employ multiple models and sequence runs across different models and even different providers.

Why is multi-model so important?

cost-optimization
Cost optimization

For some tasks and activities, some models are just more expensive than others. So it is critical that you have a choice in which models you work with and that you can seamlessly switch between models.

balance
Load balancing & failover

Vertesia is the only company to offer virtualized environments that can balance workloads across different models and providers. We also support multi-model failover, enabling you to automatically switch models or providers in the event that your current model is unavailable or is returning erroneous results.

unlocked
Prevent vendor lock-in

We abstract the underlying GenAI model from your app, service, or agent. This approach not only ensures that you are working with the optimal model but also avoids vendor lock-in. making it painless to adopt new inference models as they become available.

How do you know which GenAI model will deliver the best results?

We provide our own GenAI-driven agent in Vertesia Studio to help you quickly determine the optimal model or models for your use case.
OpenAI-600x240
Gemini-600x240
Amazon-Nova-600x240
Google-Gemma-600x240
Amazon-Titan-600x240
Claude-600x240
IBM-Granite-600x240
Cohere-600x240
Meta-600x240
AI21_Labs_600x240
Mistral-600x240
DeepSeek-600x240
Also, our Services team is here to help guide you along the way
INFERENCE PROVIDERS

Currently supported inference providers and generative AI models

Inference providers host, and sometimes build their own, generative AI (GenAI) models, including large language models (LLMs) and small language models (SLMs), and are continuously adding new models to their libraries

Vertesia immediately adds new models as soon as they become available, ensuring that our customers get instant access to the latest GenAI model advancements.

Amazon Bedrock

Amazon Bedrock provides their own foundation models (Nova & Titan) plus access to additional models from other providers. Our open-source connector to Amazon Bedrock provides access to all available foundation models within Amazon Bedrock.
Bedrock-600x240

NOTE:
Below is just a sample of available models from various model providers within the Amazon Bedrock environment in Vertesia Studio. As new models are released, they are immediately made available in Vertesia. If you have questions about our support for a particular model, please contact us.

Amazon-Nova-600x240


  • Nova Lite
  • Nova Micro
  • Nova Pro

Amazon-Titan-600x240


  • Titan Embeddings
  • Titan Multimodal Embeddings
  • Titan Text Large
  • Titan Text Embeddings V2
  • Titan Text G1 - Express
  • Titan Text G1 - Lite
  • Titan Text G1 - Premier

Anthropic-600x240


  • Claude
  • Claude Instant
  • Claude 3 Haiku
  • Claude 3 Opus
  • Claude 3 Sonnet
  • Claude 3.5 Sonnet
  • Claude 3.5 Sonnet v2

AI21_Labs_600x240


  • Jamba 1.5 Large
  • Jamba 1.5 Mini
  • Jamba Instruct
  • J2 Grande Instruct
  • J2 Jumbo Instruct
  • Jurassic-2 Mid
  • Jurassic-2 Ultra

Cohere-600x240


  • Command Light
  • Command R
  • Command

Meta-600x240


  • Llama 3 8B Instruct
  • Llama 3.1 8B Instruct
  • Llama 3.1 70B Instruct
  • Llama 3.2 1B Instruct
  • Llama 3.2 3B Instruct
  • Llama 3.2 11B Instruct
  • Llama 3.2 90B Instruct
  • Llama 3.3 70B Instruct

Mistral-600x240


  • Mistral 7B Instruct
  • Mistral Small
  • Mixtral 8x7B Instruct

Azure OpenAI

Azure OpenAI provides access to OpenAI's language models and embeddings model series. These models can be easily adapted to your specific task including but not limited to content generation, summarization, image understanding, semantic search, and natural language to code translation.
Azure-plus-OpenAI

NOTE:
Below is just a sample of the available OpenAI models within the Azure OpenAI environment in Vertesia Studio. As new models are released, they are immediately made available in Vertesia. If you have questions about our support for a particular model, please contact us.

OpenAI-600x240


  • GPT-3.5
  • GPT-4
  • GPT-4 Turbo
  • GPT-4o
  • GPT-4o mini
  • o1
  • o1 mini
  • o1 preview
  • o3 mini
  • Embeddings

Google Vertex AI

Google provides their own foundation models (Gemini, Gemma, Imagen, etc.) plus access to additional models from other providers. Our open-source connector to Google Vertex AI provides access to the following foundation models.
Vertex-600x240

NOTE:
Below is just a sample of available models from various model providers within the Google Vertex AI environment in Vertesia Studio. As new models are released, they are immediately made available in Vertesia. If you have questions about our support for a particular model, please contact us.

Gemini-600x240


  • Gemini 2.0 Flash
  • Gemini 2.0 Pro
  • Gemini 2.0 Flash-Lite
  • Gemini 2.0 Flash Thinking
  • Gemini 1.5 Flash
  • Gemini 1.5 Pro
  • Gemini 1.0 Pro
  • Gemini 1.0 Pro Vision
  • Gemini Pro
  • Gemini Pro Vision
  • Gemini Ultra Vision
  • Imagen 3

Anthropic-600x240


  • Claude 3 Haiku
  • Claude 3.5 Haiku
  • Claude 3 Opus
  • Claude 3.5 Sonnet
  • Claude 3.5 Sonnet v2

Groq

Groq is a fully integrated inference provider and execution environment in which language models run. Groq created the first Language Processing Unit™ (LPU™) Inference Engine. It is an end-to-end processing unit system that provides extremely fast inference for computationally intensive applications with a sequential component to them.
Groq-600x240

NOTE:
Below is just a sample of available models from various model providers within the Groq environment in Vertesia Studio. As new models are released, they are immediately made available in Vertesia. If you have questions about our support for a particular model, please contact us.

Meta-600x240


  • Llama 3 70B
  • Llama 3 8B
  • Llama 3.1 8B
  • Llama 3.3 70B
  • Llama Guard 3 8B
  • Llama 3.2 1B*
  • Llama 3.2 3B*
  • Llama 3.2 11B*
  • Llama 3.2 90B*

*Preview models are intended for evaluation purposes only and should not be used in production environments as they may be discontinued at short notice.

OpenAI-600x240


  • Whisper Large v3
  • Whisper Large v3 turbo

Mistral-600x240


  • Mixtral 8x7B

DeepSeek-600x240


  • DeepSeek R1 Distill Qwen 32B*
  • DeepSeek R1 Distill Llama 70B*

*Preview models are intended for evaluation purposes only and should not be used in production environments as they may be discontinued at short notice.

Hugging Face

We support Hugging Face inference endpoints. Developers can easily deploy transformers, diffusers or any model on dedicated, fully managed infrastructure.
HuggingFace-600x240

NOTE:
Below is just a sample of available models from various model providers within the Hugging Face environment in Vertesia Studio. As new models are released, they are immediately made available in Vertesia. If you have questions about our support for a particular model, please contact us.

Mistral-600x240


  • Mixtral 8x22B
  • Mistral 8x7B
  • Mistral 7B Instruct
  • Pixtral 12B
  • Mistral Small
  • Mistral Large
  • Mistral Nemo
  • Mamba Codestral
  • Codestral
  • Mathstral

DeepSeek-600x240


  • DeepSeek R1
  • DeepSeek V3
  • DeepSeek VL2
  • DeepSeek Prover
  • DeepSeek V2
  • DeepSeek Coder V2
  • DeepSeek Math
  • DeepSeek VL
  • DeepSeek Coder
  • DeepSeek LLM
  • DeepSeek MoE
  • DeepSeek V2.5
  • ESFT

Elyza-600x240


  • Japanese Llama 2 7B
  • Japanese Llama 3
  • Japanese Llama 2 13B
  • Japanese CodeLlama 7B

Google-600x240


  • PaliGemma 2
  • PaliGemma
  • PaliGemma FT
  • MetricX 23
  • MetricX 24
  • HAI-DEF
  • Gemma 2
  • Gemma 2 2B
  • CodeGemma

IBM watsonx™

The IBM watsonx™ foundation models library is available in Vertesia and provides access to all available foundational models in the library, including Granite, IBM's flagship series of LLM foundation models based on decoder-only transformer architecture.
IBM_WatsonX-600x240

NOTE:
Below is just a sample of available models from various model providers within the IBM watsonx™ foundation models library environment in Vertesia Studio. As new models are released, they are immediately made available in Vertesia. If you have questions about our support for a particular model, please contact us.

IBM-Granite-600x240


  • Granite Guardian 3 2B
  • Granite Guardian 3 8B
  • Granite 13B Chat
  • Granite 13B Instruct
  • Granite 3 2B Instruct
  • Granite 3 8B Instruct
  • Granite 20B Multilingual
  • Granite 3B Code Instruct
  • Granite 8B Code Instruct
  • Granite 20B Code Instruct
  • Granite 34B Code Instruct
  • Granite 8B Japanese
  • Granite Embedding 107M Multilingual
  • Granite Embedding 278M Multilingual
  • Slate 125M English
  • Slate 30M English

Meta-600x240


  • Llama 3.2 90B Vision Instruct
  • Llama 3.2 11B Vision Instruct
  • Llama 3.2 1B Instruct
  • Llama 3.2 3B Instruct
  • Llama 3 405B Instruct

Mistral-600x240


  • Pixtral 12B
  • Mistral Large 2
  • Mistral Small 24B Instruct
  • Mixtral 8x7B Instruct

Google-600x240


  • Flan T5 XL 3B
  • Flan T5 XXL 11B
  • Flan UL2 20B

OpenAI

OpenAI, provider of the popular chatbot, ChatGPT, provides a number of foundation models which can be accessed via our open-source connector to the OpenAI model library.
OpenAI-600x240

NOTE:
Below is just a sample of available models within the OpenAI environment in Vertesia Studio. As new models are released, they are immediately made available in Vertesia. If you have questions about our support for a particular model, please contact us.

OpenAI-600x240


  • GPT-3.5 Turbo
  • GPT-4
  • GPT-4 Turbo
  • GPT-4o
  • GPT-4o mini
  • o1
  • o1 mini
  • o1 preview
  • o3 mini
  • Embeddings
  • Moderation
  • GPT base: babbage & davinci

Mistral AI

Mistral AI is a fully integrated inference and GenAI model provider. Mistal AI provides fast, open-source and secure language models.
Mistral-600x240

NOTE:
Below is just a sample of available models within the Mistral AI environment in Vertesia Studio. As new models are released, they are immediately made available in Vertesia. If you have questions about our support for a particular model, please contact us.

Mistral-600x240


  • Codestral
  • Mistral Large
  • Pixtral Large
  • Ministral 3B
  • Ministral 8B
  • Mistral Embed
  • Mistral Moderation
  • Mistral Small
  • Pixtral
  • Mistral Nemo
  • Codestral Mamba
  • Mathstral
  • Mistral 7B
  • Mixtral 8x7B
  • Mixtral 8x22B
  • Mistral Medium

Replicate

Replicate is a fully integrated inference provider and execution environment in which language models run. Replicate lets developers run machine learning (ML) models in the cloud. Users can run open-source models that other people have published, or package and publish their own models.
Replicate-600x240

NOTE:
Below is just a sample of available models from various model providers within the Replicate environment in Vertesia Studio. As new models are released, they are immediately made available in Vertesia. If you have questions about our support for a particular model, please contact us.

Meta-600x240


  • Llama 3 70B
  • Llama 3 70B Instruct
  • Llama 3 8B Instruct
  • Llama 3 8B
  • Llama 2 7B
  • Llama 2 7B Chat
  • Llama 2 13B
  • Llama 2 13B Chat
  • Llama 2 70B Chat
  • Llama 3.1 405B Instruct
  • Llama Guard 2 8B
  • Llama Guard 3 8B
  • CodeLlama 7B
  • CodeLlama 7B Instruct
  • CodeLlama 70B Instruct
  • CodeLlama 34B
  • CodeLlama 13B
  • CodeLlama 13B Instruct
  • CodeLlama 13B Python
  • Musicgen
  • Detic

Anthropic-600x240


  • Claude 3.5 Haiku
  • Claude 3.5 Sonnet

Mistral-600x240


  • Mistral 7B
  • Mistral 7B Instruct

OpenAI-600x240


  • Whisper

Google-600x240


  • Imagen 3 Fast
  • Imagen 3

DeepSeek-600x240


  • DeepSeek R1

Together AI

Together AI is a fully integrated inference provider and execution environment in which language models run. Together AI allows developers to run open-source models like Llama2 on one of the fastest inference stacks available.
TogetherAI-600x240

NOTE:
Below is just a sample of available models from various model providers within the Together AI environment in Vertesia Studio. As new models are released, they are immediately made available in Vertesia. If you have questions about our support for a particular model, please contact us.

Meta-600x240


  • Llama 3.3 70B
  • Llama 3.2 11B
  • Llama 3.2 90B
  • Llama 3.2 3B Instruct Turbo
  • Llama 3.1 8B
  • LLama 3.1 70B
  • Llama 3.1 405B
  • Llama 2 Chat
  • Llama Guard 7B
  • Llama Guard 2 8B

Mistral-600x240


  • Mistral
  • Mistral Small
  • Mistral Instruct
  • Mistral 7B Instruct V2
  • Mistral 7B Instruct V3

DeepSeek-600x240


  • DeepSeek R1
  • DeepSeek R1 Distilled Llama 70B
  • DeepSeek V3
  • DeepSeek R1 Distilled Qwen 14B
  • DeepSeek R1 Distilled Qwen 1.5B
  • DeepSeek 67B

Google-Gemma-600x240


  • Gemma Instruct 2B
  • Gemma 2 Instruct 27B
  • Gemma 2 Instruct 9B
OPEN-SOURCE CONNECTORS

Easily connect to inference providers

In Vertesia, environments are where you connect to inference providers. Simply add your API key to connect to any of the major providers and access their GenAI foundation models using our open-source connectors.

FAQ

Frequently Asked Questions

Does Vertesia host models?

No, we do not host generative AI models. We instead provide unified access to leading inference providers who host models and to model builders who host their own models, allowing our customers to leverage hundreds of models from different providers to easily select the right model for their particular use case.

Can I use a custom model?

Of course. Vertesia enables customers to connect to their own custom models. We also provide functionality to assist customers in fine-tuning custom models for optimal performance.

What is the difference between a model builder and an inference provider?

Model builders are companies that create and train AI models, including the underlying algorithms and data sets. Inference providers are companies that provide the infrastructure and access to run the AI models. Inference providers often create and train their own models, but inference providers may also operate models from other model providers.

Inference providers that only host models include:

  • Groq
  • Microsoft (Azure)
  • Replicate
  • Together AI

Inference providers that host & build models include:

  • OpenAI
  • Google
  • Amazon
  • IBM
  • Mistral AI

Model builders that do not host their models include:

  • Meta
  • DeepSeek
  • Anthropic
  • Cohere
GET STARTED

Are you intrigued? Let's schedule a demo!

Many organizations don’t understand the full potential of GenAI or how they can transform business processes. Once they see our platform in action, they often have an "aha!" moment, realizing how we can solve real-world challenges.

See how our platform accelerates IT and business outcomes with a live, personalized demo.