TextQL Inference Options
Inference Overview
Overview
Large language models are at the core of our product, so making the way we use language models transparent and customizable for your organization is paramount.
In-Depth
There are three main types of language model deployment that we support
- Foundational Model Providers
- In-your-cloud Foundation Models
- Self-hosted Open Source Models
The best performing versions of TextQL use some variation of a closed-source foundational model to run core LLM functions. For our multi-tenant deployment this defaults to OpenAI (GPT-4o, o1-mini) as we’ve found this to have the best performance.
We’ve also found comparable performance from Anthropic’s Claude 3.5 Sonnet and thus this is also and officially supported deployment type.
Lastly we have the ability to run TextQL with publicly available open source models, though performance trails OpenAI and Anthropic and it is not recommended.
Foundational Model Provider
Overview
The fastest and easiest way to get industry leading language model performance, all without sacrificing privacy and security
In-Depth
Working directly with foundational model providers means getting access to the state of the art language models right as they’re released.
We work closely with the teams at both OpenAI and Anthropic to use the latest and greatest LLMs that they put out, all while maintaining the data security and privacy that you expect from an enterprise product. We have DPAs in place with both providers and review their SOC 2 Type 2 audits.
For OpenAI we use
- gpt-4o
- o1-mini
For Anthropic we use
- claude-3.5-sonnet
- claude-3.5-haiku
Why?
- Best in class models
- High Availability
- SOC2, GDPR, HIPAA certified/compliant
Who it’s for
- Organization looking for the best performance and pricing
In-Your-Cloud Model Providers
Overview
Use best-in-class models all while making sure data never leaves the VPC.
** We can connect to In-Your-Cloud Model Providers within your VPC from Multi and Single tenant managed deployments (via PrivateLink or internet). Contact us for more information.
In-Depth
Foundational model providers now partner with leading clouds to allow you to deploy the same best in class LLMs available on their own APIs directly within your VPC. Each of the major public clouds has partnerships with either Anthropic or OpenAI and we support both.
For AWS:
- AWS Bedrock within you VPC
- Anthropic Claude 3.5 Sonnet and Haiku
For GCP:
- GCP Vertex AI
- Anthropic Claude 3.5 Sonnet and Haiku
For Azure:
- Azure OpenAI
- OpenAI GPT-4o and o1-mini
Why
- Best in class models
- High availability
- Cloud inference priced into your cloud bill
- Data stays in your VPC
Who it’s for
- Organization with large cloud credit balances
- Organizations seeking to keep data from leaving their VPC
Self-hosted Models
Overview
For maximum security and control, run models where you know every weight and bias.
In-Depth
For organizations that require visibility into model weights and runner environments, we offer the ability to run TextQL with open source models.
This performs meaningfully worse than the models offered by OpenAI and Anthropic so we reserve this for organizations in highly regulated industries where this is required.
Why
- Full control over models
- Ability to fine tune or additionally pre-train
Who it’s for
- Organization with restrictions on LLM use