AI engines & cascade
How translation engines work
SwiftIn uses different translation engines depending on your plan. AI models give the best quality; Google and Bing handle volume and act as a free fallback.
Available engines
Grok 4.1 Fast
Free with quota / Pro / TeamX's fast model. Default for AI translation. Casual tone, good for social media context.
DeepSeek v4 Flash
Free with quota / Pro / TeamStrong on technical and Asian-language content. Switch in settings.
Gemini 2.5 Flash Lite
Free with quota / Pro / TeamGoogle's lightweight AI model. Context-aware, supports translation styles.
Google Translate
Free / GuestStandard machine translation. Reliable, fast, no API key needed. Works for guests too.
Bing Translator
Free / GuestMicrosoft's engine. Used as a fallback when Google rate-limits. Lower concurrency.
More models coming
We're testing and adding new AI engines. Stay tuned.
How fallback works
When one engine fails or hits a limit, SwiftIn falls back automatically — you don't hit an error wall.
Guest path
google → google-anon → bing. Each step retried on failure.
Paid AI path
Selected AI model → other AI models → Google → Bing. Failures cascade through cheaper options.
When AI quota runs out
Free users with 100K AI tokens used: AI engines silently downgrade to Google + Bing. Pro/Team get larger quotas (5M / 10M per seat).
Switching engines
Pick a per-feature engine in extension settings: Page translate, Selection, Input each have their own engine setting. Your choice persists across sessions.
Cache & token deduction
Translations are cached locally for 30 days. Revisits don't spend tokens.
Local translation cache
Each translation is saved in your browser's IndexedDB for 30 days. When you revisit a page, the translation loads from cache instantly — no engine call, no token deduction.
Free engines = no limits
Google + Bing run without tokens and without limits. Translate any pages, as many times as you want — never charged, never throttled.
Paid AI = first request bills, then cache
Gemini, DeepSeek, Grok deduct tokens only on the first translation. Visit the same page later — translation loads from your local cache, tokens stay untouched.
Switching engine = fresh request
If you translated a page with Gemini and switch to DeepSeek, DeepSeek makes a fresh request (cache is keyed per engine). Switch to a paid engine — tokens deduct. Switch to Google/Bing — no charge.
AI Smart Context
AI Smart Context tells the AI engine what the page is about before it translates. Result: tone, jargon, and references match the actual content — not just the words.
Why context matters
Without context, AI translates each paragraph in isolation. Technical articles get casual translations; casual chats get formal ones. With context, the model knows the page is a finance article, a meme thread, or a docs page — and adjusts.
What's used as context
Page title (always, all plans), meta description (when present and substantial), and an AI-generated summary (~300–500 tokens, Pro/Team only).
Summary cache (24h)
The summary is generated once per URL + first 500 chars of content, then cached for 24 hours. Other Pro/Team users hitting the same page in that window get the cached summary for free.
Token cost
Each summary costs ~300–500 tokens of your AI quota — only on the page's first translation. Subsequent translations (within the 24h window) reuse the cached summary at no cost.
Free vs Pro/Team
Free: title-only context (no AI summary, no token cost). Pro/Team: full Title + Meta + AI summary for context-aware translations.
Opt-in by default
Off until you turn it on. Toggle in: extension popup → Translate card → Smart Context, or in Options → Page Translate tab.