The short version. BYOK means "bring your own key". You sign up for an API key with Anthropic, OpenAI, or Google. You paste it into a desktop app. The app uses your key to talk to the AI, and you pay the AI provider directly for what you used. The app charges you separately, often a one-time fee. For most indie founders this works out to $2-15 per month in AI costs, which is lower than any flat $20/month subscription tool, with better privacy guarantees.
I've spent a lot of time the last year explaining BYOK to founders who were stacking $20/month subscriptions and wondering why their tooling bill kept growing. The short answer is that they were paying a margin on AI inference twice: once to the tool that bundled it, and once more in the form of a vendor lock-in tax that compounded as their data piled up.
BYOK fixes both of those, with one real trade-off (you have to set up an API key once). This page walks through what BYOK actually means, what it costs in practice, why it changes the privacy math, and how to set up your first key in about five minutes.

Most AI tools come in one of two shapes:
The acronym BYOK shows up in the cloud-software world ("bring your own key") to mean a customer-managed encryption key. In the AI tools context, it means the same idea applied to your AI provider account: you're the customer of the AI provider, not of the tool wrapping it.
It matters because it inverts who owns the relationship. With bundled inference, you're a customer of Notion / OpenAI / Anthropic-via-the-tool. With BYOK, you're a customer of Anthropic / OpenAI / Google directly, and the tool is just an interface.
Here's what BYOK actually costs in practice, from real usage logs.
Typical token usage: ~80,000 tokens/day in, ~40,000 out. With Claude Sonnet 4.5 ($3 per million input, $15 per million output), that's:
With OpenAI's GPT-5 mini or Google's Gemini 2.5 Flash, both significantly cheaper for similar quality on most tasks, the same volume runs $5-12/month.
~250,000 tokens/day in, ~120,000 out. Same Claude Sonnet 4.5 pricing:
Heavy users on the cheaper Gemini Flash or GPT-mini tier: $15-30/month.
| Option | Light user / mo | Heavy user / mo | 3-year total (light) |
|---|---|---|---|
| ChatGPT Plus | $20 | $20 (with throttling) | $720 |
| Claude Pro | $20 | $20 (with throttling) | $720 |
| Notion + Notion AI | $10 + $10 = $20 | $10 + $10 = $20 | $720 |
| Cursor Pro | $20 | $20 (with throttling) | $720 |
| Projelli (Pro $49) + BYOK Claude | $49 + $5-15 = year 1: $109; ongoing: $5-15/mo | $49 + $30-50 = year 1: $409; ongoing: $30-50/mo | ~$229 (light user) |
| Projelli (Lifetime $99) + BYOK Gemini | $99 + $3-8 = year 1: $135; ongoing: $3-8/mo | $99 + $15-30 = year 1: $279; ongoing: $15-30/mo | ~$252 (light user) |
The flat-rate tools win in one specific case: you genuinely use AI for many hours a day, every day, without throttling. For most indie founders, BYOK comes out cheaper across any reasonable usage pattern.
Important caveat: usage varies. The numbers above are realistic for a founder who does most of their planning, drafting, and review with AI assistance. Lower-touch users can run as low as $1-3/month. Higher-touch users (running long workflows daily, processing customer interview transcripts) can hit $50+/month.
The cost angle is what gets attention. The privacy angle is more important and less talked about.
Three things change when you switch from a bundled-inference consumer tool to BYOK on the same AI provider's API tier.
Consumer tier (ChatGPT.com, Claude.ai, Gemini app) and API tier (BYOK) have different default policies, even though the same company runs both. In short:
The takeaway: API tier is generally a stricter privacy posture than the consumer chatbot. Confirm the current terms before relying on this; they change.
With a bundled-inference tool, your prompt goes: your machine → tool's servers → AI provider → tool's servers → your machine. The tool's company sees your prompt and the response.
With a BYOK desktop app like Projelli, your prompt goes: your machine → AI provider → your machine. The desktop app's servers are not in the path. There's no intermediate store. Cursor, Continue.dev, Projelli, and most BYOK tools work this way. (Always check; some tools route through their own servers even with BYOK, which is a different threat model.)
Even if the AI provider has a strict no-training policy, the question of where your conversation lives still matters. With a bundled tool, the conversation is in their database. With a BYOK desktop app, the conversation is a file on your hard drive, in a folder you picked.
For an indie founder whose AI conversations contain pitch decks, customer interviews, and pricing strategy work, this is the difference between "in someone else's database with a TOS that can change" and "on my machine, period". Most founders prefer the second posture once they think about it.
Five minutes, three providers. Pick one to start; you can add the others later.
sk-ant-)You're done. Claude conversations now go directly from your machine to Anthropic.
sk-)AIza)Gemini has a generous free tier; you may not need to enter a payment method to start. Read the data terms before relying on the free tier for sensitive content.
All three providers let you set a monthly spend cap. Use it. A typical sequence the first time:
The cap is a hard ceiling. Once you hit it, the API stops responding until next billing cycle. This protects you from runaway costs if a tool starts looping (rare but possible).
You can run Llama, Mistral, or other open-weight models locally with tools like Ollama or LM Studio. This is fully off-cloud: no API key, no provider, no internet required. Privacy maxes out, cost is zero (except electricity), and the data stays entirely on your machine.
The trade-off is quality. Open-weight models are catching up but are not at parity with Claude Sonnet, GPT-5, or Gemini 2.5 Pro for the kind of strategic-thinking work most indie founders want AI for. Drafting code? Open-weight models are competitive. Writing a pitch deck? You'll feel the difference.
The realistic 2026 answer: use BYOK with Claude/OpenAI/Gemini for high-stakes work, and use local Ollama for the volume work where 90% quality is fine. Projelli supports both: configure local Ollama as a provider alongside Claude / OpenAI / Gemini, and route per-conversation.
Some people will. They're not the target market for BYOK tools. The audience that buys local-first BYOK apps is exactly the audience comfortable getting an API key. If "5 minutes setting up an account at console.anthropic.com" is a dealbreaker, the bundled-inference tools are a fine choice.
Google's Gemini has a generous free tier that doesn't always require a card. Anthropic and OpenAI both require payment info to use the API at all, even with a $5 prepaid balance. If avoiding cards is your priority, Gemini is the right starting point. After a month of using it for free, you'll have a sense of what your actual usage is, and adding a card to one of the others becomes a smaller mental ask.
Three layers of protection. First, the desktop app stores the key in your OS keychain (macOS) or an encrypted file (Windows / Linux), not in plaintext anywhere. Second, all three providers let you revoke a key with one click and generate a new one. Third, the monthly spend cap means even a stolen key has bounded damage. Treat the key like a password: don't share it, rotate if you suspect exposure, and you're fine.
Same risk as a bundled-inference tool, since they use the same providers under the hood. A BYOK app actually gives you a small advantage: you can quickly switch providers (Claude → OpenAI → Gemini) if one is having an outage. Bundled tools are stuck with whatever they wrap.
| Tool | BYOK support | Workspace shape |
|---|---|---|
| Projelli | Yes, all 3 providers | Local-first AI workspace for founders |
| Cursor | Yes (Pro tier) | Code editor with AI |
| Continue.dev | Yes | VS Code AI extension |
| Raycast AI | Yes (Advanced) | Mac launcher with AI |
| Obsidian + Smart Connections / Copilot | Yes | Local notes with AI plugins |
| ChatGPT, Claude.ai, Gemini app | No | Cloud chat, bundled inference |
| Notion AI | No | Cloud workspace, bundled inference |
| Reflect, Mem.ai, Tana | No | Cloud notes, bundled inference |
This list moves. Some tools are adding BYOK as a Pro option; some are removing it. Check the current state of any tool you're evaluating.
BYOK stands for "bring your own key". You sign up for an API key with an AI provider (Anthropic, OpenAI, or Google), paste it into the desktop app, and the app uses your key to make AI requests. You pay the AI provider directly for usage; the desktop app charges you separately or one-time, but never marks up inference.
For a typical indie founder doing 30-90 minutes of AI work per day, BYOK costs $2-15 per month with Claude, OpenAI, or Gemini. That's lower than the $20/month flat for ChatGPT Plus, Claude Pro, or Notion AI. Heavy users (multi-hour daily, large context) pay more, $20-40/month, but still typically less than stacked subscriptions.
Yes, in two specific ways. First, AI providers' API tier (which is what BYOK uses) typically does not train on your data by default, while consumer tiers often do. Second, with a BYOK desktop app, the app's company never sees your prompts; the requests go directly from your machine to the AI provider.
Anthropic Claude: console.anthropic.com → Settings → API Keys. OpenAI: platform.openai.com → API Keys. Google Gemini: aistudio.google.com → Get API key. Each takes 2-3 minutes to sign up and generate. All three require entering payment info to start, but the free tiers and prepaid balances are real (you can put $5 in and use it down).
All three providers let you set a hard monthly spending cap. If you set a $20 cap, the API will stop responding when you hit it instead of charging more. This protects you from runaway costs while you learn how much you actually use.
Most desktop AI workspaces designed for power users support BYOK: Projelli, Cursor (for code), Continue.dev (for code), several Obsidian AI plugins, Raycast AI. Most cloud workspaces (Notion AI, ChatGPT Plus, Claude.ai Pro, Reflect, Mem.ai) do not.
If the tool supports multiple providers, yes. Projelli ships with all three (Claude, OpenAI, Gemini) plus an Ollama integration for local models. You add keys for each, and pick which provider to use on a per-conversation basis. The advantage: if one provider is down or rate-limited, you switch in one click.
Ollama and LM Studio let you run open-weight models entirely on your machine. No API key, no internet, fully offline. Quality is competitive for code and short-form tasks; it lags behind Claude / GPT / Gemini for long-form strategic work. Most founders use a hybrid: BYOK Claude for high-stakes work, local Ollama for volume.
Free download. $49 Pro one-time. $99 Lifetime. Founder's tier is $29 for the first 100 buyers. Bring your own Claude / OpenAI / Gemini key.
Get Projelli