Skip to content

AI engines & cascade

How translation engines work

SwiftIn uses different translation engines depending on your plan. AI models give the best quality; Google and Bing handle volume and act as a free fallback.

Available engines

Grok 4.1 Fast

Free with quota / Pro / Team

X's fast model. Default for AI translation. Casual tone, good for social media context.

DeepSeek v4 Flash

Free with quota / Pro / Team

Strong on technical and Asian-language content. Switch in settings.

Gemini 2.5 Flash Lite

Free with quota / Pro / Team

Google's lightweight AI model. Context-aware, supports translation styles.

Google Translate

Free / Guest

Standard machine translation. Reliable, fast, no API key needed. Works for guests too.

Bing Translator

Free / Guest

Microsoft's engine. Used as a fallback when Google rate-limits. Lower concurrency.

More models coming

We're testing and adding new AI engines. Stay tuned.

How fallback works

When one engine fails or hits a limit, SwiftIn falls back automatically — you don't hit an error wall.

Guest path

google → google-anon → bing. Each step retried on failure.

Paid AI path

Selected AI model → other AI models → Google → Bing. Failures cascade through cheaper options.

When AI quota runs out

Free users with 100K AI tokens used: AI engines silently downgrade to Google + Bing. Pro/Team get larger quotas (5M / 10M per seat).

Switching engines

Pick a per-feature engine in extension settings: Page translate, Selection, Input each have their own engine setting. Your choice persists across sessions.

Cache & token deduction

Translations are cached locally for 30 days. Revisits don't spend tokens.

Local translation cache

Each translation is saved in your browser's IndexedDB for 30 days. When you revisit a page, the translation loads from cache instantly — no engine call, no token deduction.

Free engines = no limits

Google + Bing run without tokens and without limits. Translate any pages, as many times as you want — never charged, never throttled.

Paid AI = first request bills, then cache

Gemini, DeepSeek, Grok deduct tokens only on the first translation. Visit the same page later — translation loads from your local cache, tokens stay untouched.

Switching engine = fresh request

If you translated a page with Gemini and switch to DeepSeek, DeepSeek makes a fresh request (cache is keyed per engine). Switch to a paid engine — tokens deduct. Switch to Google/Bing — no charge.

AI Smart Context

AI Smart Context tells the AI engine what the page is about before it translates. Result: tone, jargon, and references match the actual content — not just the words.

Pro/Team only. Free users with the toggle on get title-only context (no AI summary).

Why context matters

Without context, AI translates each paragraph in isolation. Technical articles get casual translations; casual chats get formal ones. With context, the model knows the page is a finance article, a meme thread, or a docs page — and adjusts.

What's used as context

Page title (always, all plans), meta description (when present and substantial), and an AI-generated summary (~300–500 tokens, Pro/Team only).

Summary cache (24h)

The summary is generated once per URL + first 500 chars of content, then cached for 24 hours. Other Pro/Team users hitting the same page in that window get the cached summary for free.

Token cost

Each summary costs ~300–500 tokens of your AI quota — only on the page's first translation. Subsequent translations (within the 24h window) reuse the cached summary at no cost.

Free vs Pro/Team

Free: title-only context (no AI summary, no token cost). Pro/Team: full Title + Meta + AI summary for context-aware translations.

Opt-in by default

Off until you turn it on. Toggle in: extension popup → Translate card → Smart Context, or in Options → Page Translate tab.

Translation stylesText-to-Speech