Choosing the Right AI Model for Your Private Deployment

One of the most important decisions in a RowSpeak Private Deployment is which AI model to use. The right choice depends on your organization's privacy requirements, infrastructure, performance needs, and internal policy.

This guide walks through the options and helps you decide.


The Two Paths

Open-Source vs. Closed-Source Models Open-Source Models Run entirely inside your environment Examples: Qwen3.5 · DeepSeek V3 · Gemma 4 Data handling: ✅ Data never leaves your network ✅ No external API calls ✅ Air-gapped compatible Requirements: GPU server (16–64 GB VRAM) Closed-Source Models Via your own API key Examples: GPT-5.4 · Claude Opus 4.6 · Gemini 3.1 Pro Data handling: ⚠️ Data sent to provider API ⚠️ Subject to provider's policies ✅ Your API key, your account Requirements: API key + outbound internet access

Decision Framework

Use this to narrow down your choice quickly.

Choose open-source if:

  • Your data cannot leave your network under any circumstances
  • You are deploying in an air-gapped environment
  • You have regulatory or compliance requirements around data residency
  • You want to avoid dependency on third-party API availability
  • You want predictable costs without per-token pricing

Choose closed-source if:

  • Your data sensitivity allows external API calls (with your own account)
  • You want the highest possible output quality for complex reasoning tasks
  • You do not have GPU infrastructure available
  • You want to get started quickly without model setup
  • You are running a pilot before committing to GPU hardware

Mix both if:

  • Different workflows have different sensitivity levels
  • You want open-source for sensitive data and closed-source for non-sensitive tasks
  • You want a fallback model if one is unavailable

RowSpeak supports routing different workflows to different models within the same deployment.


Open-Source Model Comparison

Model Parameters VRAM Required Languages License Best For
Qwen3.5-397B-A17B 397B (MoE, 17B active) 64 GB 201+ languages Apache 2.0 Maximum quality, complex analysis
Qwen3.5-122B-A10B 122B (MoE, 10B active) 24–48 GB 201+ languages Apache 2.0 High-quality, complex tasks
Qwen3.5-35B-A3B 35B (MoE, 3B active) 16 GB 201+ languages Apache 2.0 Multilingual, general analysis
DeepSeek V3 671B (MoE, 37B active) 32–48 GB EN, ZH, multilingual MIT Math reasoning, code generation
Gemma 4-31B 31B 24 GB 140 languages Apache 2.0 Agentic workflows, structured output
Qwen3.5-9B 9B 16 GB 201+ languages Apache 2.0 Cost-effective local deployment

All open-source models run locally. No internet connection required after initial setup. Licenses allow commercial use.


Closed-Source Model Comparison

Model Provider Context Window Best For
GPT-5.4 OpenAI 1M tokens Complex reasoning, English-first
Claude Opus 4.6 Anthropic 1M tokens Long documents, nuanced analysis
Gemini 3.1 Pro Google 1M tokens Very large files, mixed content

You use your own API key. RowSpeak does not act as an intermediary — your data goes directly from your environment to the provider.


What Happens to Your Data with Closed-Source Models

When you use a closed-source model, your spreadsheet data is sent to that provider's API for processing. This means:

  • The provider processes your data according to their terms of service
  • Enterprise API agreements with OpenAI, Anthropic, and Google typically include data processing agreements (DPAs) that restrict training use
  • You should review the provider's data handling policies before enabling this option
  • RowSpeak recommends using closed-source models only for non-sensitive data, or after reviewing and accepting the provider's enterprise data terms

For maximum data sovereignty, use open-source models.


Model Routing in RowSpeak

RowSpeak supports configuring different models for different use cases within the same deployment.

Example configuration:

Workflow: Financial reports → DeepSeek V3 (local, sensitive data)
Workflow: Marketing summaries → GPT-5.4 (API, non-sensitive)
Workflow: Default → Qwen3.5-35B (local, general use)

This lets your organization apply the right model to each workflow based on data sensitivity, without forcing a single choice across all use cases.


Frequently Asked Questions

Can I switch models after deployment?
Yes. Model selection is a configuration change, not a re-deployment. Your IT team can update the model routing configuration without downtime.

Do I need to download model weights myself?
No. The RowSpeak Deployment Pack includes guidance on model acquisition. For air-gapped environments, we provide instructions for pre-loading model weights before deployment.

What if I want to use a model not on this list?
Contact us. RowSpeak's model layer is designed to be extensible. If you have a specific model requirement, we can discuss compatibility.

Can I use a fine-tuned or custom model?
This is available on the Enterprise tier. Contact us to discuss your requirements.


Need Help Deciding?

Book a demo and we will help you map the right model strategy to your environment, data sensitivity, and performance requirements.

You can also review the technical architecture document for more detail on how the model layer integrates with the rest of the system.