
Inference Overview
Overview
Large language models are at the core of our product, so making the way we use language models transparent and customizable for your organization is paramount.
In-Depth
There are three main types of language model deployment that we support- Foundational Model Providers
- In-your-cloud Foundation Models
- Self-hosted Open Source Models
Foundational Model Provider
Overview
The fastest and easiest way to get industry leading language model performance, all without sacrificing privacy and security
In-Depth
Working directly with foundational model providers means getting access to the state of the art language models right as they’re released. We work closely with the teams at both OpenAI and Anthropic to use the latest and greatest LLMs that they put out, all while maintaining the data security and privacy that you expect from an enterprise product. We have DPAs in place with both providers and review their SOC 2 Type 2 audits. For OpenAI we use- gpt-4o
- o1-mini
- claude-3.5-sonnet
- claude-3.5-haiku
Why?
- Best in class models
- High Availability
- SOC2, GDPR, HIPAA certified/compliant
Who it’s for
- Organization looking for the best performance and pricing
In-Your-Cloud Model Providers
Overview
Use best-in-class models all while making sure data never leaves the VPC.
In-Depth
Foundational model providers now partner with leading clouds to allow you to deploy the same best in class LLMs available on their own APIs directly within your VPC. Each of the major public clouds has partnerships with either Anthropic or OpenAI and we support both. For AWS:- AWS Bedrock within you VPC
- Anthropic Claude 3.5 Sonnet and Haiku
- GCP Vertex AI
- Anthropic Claude 3.5 Sonnet and Haiku
- Azure OpenAI
- OpenAI GPT-4o and o1-mini
Why
- Best in class models
- High availability
- Cloud inference priced into your cloud bill
- Data stays in your VPC
Who it’s for
- Organization with large cloud credit balances
- Organizations seeking to keep data from leaving their VPC
Self-hosted Models
Overview
For maximum security and control, run models where you know every weight and bias.
In-Depth
For organizations that require visibility into model weights and runner environments, we offer the ability to run TextQL with open source models. This performs meaningfully worse than the models offered by OpenAI and Anthropic so we reserve this for organizations in highly regulated industries where this is required.Why
- Full control over models
- Ability to fine tune or additionally pre-train
Who it’s for
- Organization with restrictions on LLM use