Skip to main content
Available to users on the new pricing plans (currently not available on legacy plans)

Overview

Zencoder now has a model selector in the chat input. Use the dropdown to pick which LLM runs your messages. The options you see depend on your plan and entitlements.
  • Currently, the model selector is available to users on the new pricing plans ( not available on legacy plans)
  • Model choices may differ by plan; higher plans unlock additional models
  • Select the model that fits your task from the selector (Auto, Auto+, Haiku 4.5 Parallel Thinking, Sonnet 4, Sonnet 4.5 Parallel Thinking, Opus 4.1, Opus 4.5 Parallel Thinking, Opus 4.6, Gemini Pro 3.0, GPT-5.1-Codex, GPT-5.1-Codex-mini, Grok Code Fast 1).
Zencoder supports models from OpenAI, Anthropic, Google, and xAI.

Available models (subject to change)

The Auto model routes across a tuned mix of self-hosted and vendor models for the best balance of speed, quality, and cost. Model selector showing available models and multipliers
ModelProviderMultiplierPlan Requirements
AutoZencoderAll plans
Auto+Zencoder2.5×All plans
Sonnet 4AnthropicAll plans
Haiku 4.5 Parallel ThinkingAnthropicStarter, Core, Advanced, Max
Sonnet 4.5 Parallel Thinking***AnthropicStarter, Core, Advanced, Max
Opus 4.1**Anthropic10×Advanced, Max
Opus 4.5 Parallel Thinking**AnthropicAdvanced, Max
Opus 4.6**AnthropicAdvanced, Max
Gemini Pro 3.0GoogleAll plans
GPT-5.1-CodexOpenAIStarter, Core, Advanced, Max
GPT-5.1-Codex-miniOpenAI0.5×Starter, Core, Advanced, Max
Grok Code Fast 1xAI0.25×All plans
***Sonnet 4.5 Parallel Thinking — available since September 29, 2025. Supports persistent state tracking and parallel execution, suited for spec-driven development workflows.**Opus 4.5 Parallel Thinking — available on Advanced and Max plans at 5× multiplier.*GPT-5.1-Codex — standard code generation model. GPT-5.1-Codex-mini — 0.5× multiplier, lower cost option for lighter workloads.
Lowest multipliers: Grok Code Fast 1 (0.25×), GPT-5.1-Codex-mini (0.5×), Auto (1×). All multipliers and model availability are subject to change. Model availability evolves over time. We may add, remove, or upgrade models without notice as providers change capabilities and pricing. Your selector will always reflect the currently supported set for your plan.

Cost multipliers and Premium LLM calls

Zencoder measures usage in Premium LLM calls. Some models use a multiplier to reflect provider costs or parallel reasoning. See the table above and the screenshot for current multipliers.
Example: If a response would normally consume 10 Premium LLM calls, running it with Opus 4.1 (10×) would count as 100 calls. For details about daily limits and accounting, see Pricing & Plans.

Plan-based availability

Not all plans have access to all models. The selector shows what your current plan can use. If you recently upgraded but don’t see expected options, restart your IDE to refresh entitlements.
  • Starter, Core, Advanced and Max plans: Model model selector with allocated models, including GPT-5 access
  • Legacy plans: Model selector is not available

Bring Your Own Key (BYOK)

You can use your own API key for certain providers. This is useful if you:
  • Want to remove daily Premium LLM call limits for those requests
  • Prefer billing directly with the provider
Supported providers in the UI today: Anthropic and OpenAI.
BYOK is currently available on all plans, including Free, Starter, Core, Advanced, and Max.

Enable BYOK

1

Open Zencoder menu

Click the Zencoder icon in your IDE, then click the ⋯ (three dots) menu.
2

Choose User API key

Select User API key from the dropdown to open Custom API keys.
3

Enter keys and use

Paste your Anthropic or OpenAI key, save, then pick a matching provider model in the model selector.Custom API keys screen with Anthropic and OpenAI fields
When BYOK is active for a given provider/model, those requests can bypass Zencoder’s daily Premium LLM call limits - usage and billing are governed by your provider account. Make sure you understand the provider’s pricing before enabling heavy workloads.
Keep your keys secure. Only paste keys into the official Zencoder UI. If you rotate/regenerate keys at the provider, update them in Zencoder as well.

When to switch models

  • Use the Auto model for most coding tasks - it balances speed, quality, and cost
  • Use the Auto+ for superior performance on complex tasks (2.5× multiplier)
  • Use the Haiku 4.5 Parallel Thinking if you want speed and cost-efficient option (1× multiplier)
  • Use Grok Code Fast 1 when you need the most cost-efficient option (0.25× multiplier)
  • Use GPT-5.1-Codex-mini for cost-efficient code generation (0.5× multiplier)
  • Use Gemini Pro 3.0 for high-quality performance with balanced capability (2× multiplier)
  • Use Sonnet 4.5 Parallel Thinking for spec-driven development tasks requiring persistent state tracking and parallel execution
  • Use GPT-5.1-Codex for specialized code generation tasks
  • Use Opus 4.1 for challenging tasks requiring high capability (Advanced and Max plans, 10× multiplier)
  • Use Opus 4.5 Parallel Thinking (Advanced and Max plans) for the most complex reasoning tasks, keeping the 5× multiplier in mind
  • Use Opus 4.6 (Advanced and Max plans) for enhanced reasoning capabilities with the 6× multiplier
Heavier reasoning models may produce better results on difficult tasks but at higher call consumption. Keep separate chats for different experiments to maintain clarity.

Troubleshooting

Make sure you’re on the new plans. Sign out/in to refresh entitlements.
Restart your IDE after signing back in.
Confirm you selected a model that matches the provider whose key you added, and that the key is valid/active.

Pricing & Plans

Learn how Premium LLM calls and daily limits work

Coding Agent

See how models are applied in the full agentic workflow