BYOK AI explained: bring your own AI key, the why and how.

Updated 2026-04-27 · 11 min read

The short version. BYOK means "bring your own key". You sign up for an API key with Anthropic, OpenAI, or Google. You paste it into a desktop app. The app uses your key to talk to the AI, and you pay the AI provider directly for what you used. The app charges you separately, often a one-time fee. For most indie founders this works out to $2-15 per month in AI costs, which is lower than any flat $20/month subscription tool, with better privacy guarantees.

I've spent a lot of time the last year explaining BYOK to founders who were stacking $20/month subscriptions and wondering why their tooling bill kept growing. The short answer is that they were paying a margin on AI inference twice: once to the tool that bundled it, and once more in the form of a vendor lock-in tax that compounded as their data piled up.

BYOK fixes both of those, with one real trade-off (you have to set up an API key once). This page walks through what BYOK actually means, what it costs in practice, why it changes the privacy math, and how to set up your first key in about five minutes.

Projelli Settings → API Keys: paste your Anthropic, OpenAI, or Gemini key, stored in your OS keychain

What BYOK actually is

Most AI tools come in one of two shapes:

The acronym BYOK shows up in the cloud-software world ("bring your own key") to mean a customer-managed encryption key. In the AI tools context, it means the same idea applied to your AI provider account: you're the customer of the AI provider, not of the tool wrapping it.

It matters because it inverts who owns the relationship. With bundled inference, you're a customer of Notion / OpenAI / Anthropic-via-the-tool. With BYOK, you're a customer of Anthropic / OpenAI / Google directly, and the tool is just an interface.

The cost math

Here's what BYOK actually costs in practice, from real usage logs.

Light user: 30 minutes/day of AI conversations, ~5 messages

Typical token usage: ~80,000 tokens/day in, ~40,000 out. With Claude Sonnet 4.5 ($3 per million input, $15 per million output), that's:

With OpenAI's GPT-5 mini or Google's Gemini 2.5 Flash, both significantly cheaper for similar quality on most tasks, the same volume runs $5-12/month.

Heavy user: 2 hours/day of AI work, longer context

~250,000 tokens/day in, ~120,000 out. Same Claude Sonnet 4.5 pricing:

Heavy users on the cheaper Gemini Flash or GPT-mini tier: $15-30/month.

Comparison to flat-rate alternatives

Option Light user / mo Heavy user / mo 3-year total (light)
ChatGPT Plus $20 $20 (with throttling) $720
Claude Pro $20 $20 (with throttling) $720
Notion + Notion AI $10 + $10 = $20 $10 + $10 = $20 $720
Cursor Pro $20 $20 (with throttling) $720
Projelli (Pro $49) + BYOK Claude $49 + $5-15 = year 1: $109; ongoing: $5-15/mo $49 + $30-50 = year 1: $409; ongoing: $30-50/mo ~$229 (light user)
Projelli (Lifetime $99) + BYOK Gemini $99 + $3-8 = year 1: $135; ongoing: $3-8/mo $99 + $15-30 = year 1: $279; ongoing: $15-30/mo ~$252 (light user)

The flat-rate tools win in one specific case: you genuinely use AI for many hours a day, every day, without throttling. For most indie founders, BYOK comes out cheaper across any reasonable usage pattern.

Important caveat: usage varies. The numbers above are realistic for a founder who does most of their planning, drafting, and review with AI assistance. Lower-touch users can run as low as $1-3/month. Higher-touch users (running long workflows daily, processing customer interview transcripts) can hit $50+/month.

The privacy story

The cost angle is what gets attention. The privacy angle is more important and less talked about.

Three things change when you switch from a bundled-inference consumer tool to BYOK on the same AI provider's API tier.

1. The AI provider treats your data differently

Consumer tier (ChatGPT.com, Claude.ai, Gemini app) and API tier (BYOK) have different default policies, even though the same company runs both. In short:

The takeaway: API tier is generally a stricter privacy posture than the consumer chatbot. Confirm the current terms before relying on this; they change.

2. The desktop app never sees your prompts

With a bundled-inference tool, your prompt goes: your machine → tool's servers → AI provider → tool's servers → your machine. The tool's company sees your prompt and the response.

With a BYOK desktop app like Projelli, your prompt goes: your machine → AI provider → your machine. The desktop app's servers are not in the path. There's no intermediate store. Cursor, Continue.dev, Projelli, and most BYOK tools work this way. (Always check; some tools route through their own servers even with BYOK, which is a different threat model.)

3. Your conversation history is on your machine

Even if the AI provider has a strict no-training policy, the question of where your conversation lives still matters. With a bundled tool, the conversation is in their database. With a BYOK desktop app, the conversation is a file on your hard drive, in a folder you picked.

For an indie founder whose AI conversations contain pitch decks, customer interviews, and pricing strategy work, this is the difference between "in someone else's database with a TOS that can change" and "on my machine, period". Most founders prefer the second posture once they think about it.

How to actually set up an API key

Five minutes, three providers. Pick one to start; you can add the others later.

Anthropic Claude

  1. Go to console.anthropic.com and sign up
  2. Add a payment method and put $5-20 of prepaid balance (you can set a hard monthly cap so you don't accidentally spend more)
  3. Go to Settings → API Keys → Create Key
  4. Copy the key (starts with sk-ant-)
  5. Paste it into your BYOK app (in Projelli: Settings → AI → API Keys → Anthropic)

You're done. Claude conversations now go directly from your machine to Anthropic.

OpenAI

  1. Go to platform.openai.com and sign up
  2. Add a payment method, put credit on file, and set a monthly usage cap in Settings → Limits
  3. Go to API Keys → Create new secret key
  4. Copy the key (starts with sk-)
  5. Paste into your app's OpenAI settings

Google Gemini (via AI Studio)

  1. Go to aistudio.google.com
  2. Click "Get API key" in the left nav
  3. Create or pick a Google Cloud project (the prompt is straightforward)
  4. Copy the key (starts with AIza)
  5. Paste into your app's Gemini settings

Gemini has a generous free tier; you may not need to enter a payment method to start. Read the data terms before relying on the free tier for sensitive content.

Setting a hard monthly cap (do this immediately)

All three providers let you set a monthly spend cap. Use it. A typical sequence the first time:

  1. Set the cap at $10/month while you learn your usage
  2. Use the tool for a week, see what your actual daily cost is
  3. Raise the cap to 2-3x your observed monthly usage as a safety margin

The cap is a hard ceiling. Once you hit it, the API stops responding until next billing cycle. This protects you from runaway costs if a tool starts looping (rare but possible).

What about local AI?

You can run Llama, Mistral, or other open-weight models locally with tools like Ollama or LM Studio. This is fully off-cloud: no API key, no provider, no internet required. Privacy maxes out, cost is zero (except electricity), and the data stays entirely on your machine.

The trade-off is quality. Open-weight models are catching up but are not at parity with Claude Sonnet, GPT-5, or Gemini 2.5 Pro for the kind of strategic-thinking work most indie founders want AI for. Drafting code? Open-weight models are competitive. Writing a pitch deck? You'll feel the difference.

The realistic 2026 answer: use BYOK with Claude/OpenAI/Gemini for high-stakes work, and use local Ollama for the volume work where 90% quality is fine. Projelli supports both: configure local Ollama as a provider alongside Claude / OpenAI / Gemini, and route per-conversation.

Common concerns answered

"The 5-minute setup is going to be the part everyone gives up on"

Some people will. They're not the target market for BYOK tools. The audience that buys local-first BYOK apps is exactly the audience comfortable getting an API key. If "5 minutes setting up an account at console.anthropic.com" is a dealbreaker, the bundled-inference tools are a fine choice.

"What if I don't want to attach a credit card?"

Google's Gemini has a generous free tier that doesn't always require a card. Anthropic and OpenAI both require payment info to use the API at all, even with a $5 prepaid balance. If avoiding cards is your priority, Gemini is the right starting point. After a month of using it for free, you'll have a sense of what your actual usage is, and adding a card to one of the others becomes a smaller mental ask.

"What if the API key gets stolen?"

Three layers of protection. First, the desktop app stores the key in your OS keychain (macOS) or an encrypted file (Windows / Linux), not in plaintext anywhere. Second, all three providers let you revoke a key with one click and generate a new one. Third, the monthly spend cap means even a stolen key has bounded damage. Treat the key like a password: don't share it, rotate if you suspect exposure, and you're fine.

"What if the AI provider goes down?"

Same risk as a bundled-inference tool, since they use the same providers under the hood. A BYOK app actually gives you a small advantage: you can quickly switch providers (Claude → OpenAI → Gemini) if one is having an outage. Bundled tools are stuck with whatever they wrap.

Tools that support BYOK today

Tool BYOK support Workspace shape
Projelli Yes, all 3 providers Local-first AI workspace for founders
Cursor Yes (Pro tier) Code editor with AI
Continue.dev Yes VS Code AI extension
Raycast AI Yes (Advanced) Mac launcher with AI
Obsidian + Smart Connections / Copilot Yes Local notes with AI plugins
ChatGPT, Claude.ai, Gemini app No Cloud chat, bundled inference
Notion AI No Cloud workspace, bundled inference
Reflect, Mem.ai, Tana No Cloud notes, bundled inference

This list moves. Some tools are adding BYOK as a Pro option; some are removing it. Check the current state of any tool you're evaluating.

Frequently asked questions

What does BYOK mean for AI tools?

BYOK stands for "bring your own key". You sign up for an API key with an AI provider (Anthropic, OpenAI, or Google), paste it into the desktop app, and the app uses your key to make AI requests. You pay the AI provider directly for usage; the desktop app charges you separately or one-time, but never marks up inference.

How much does BYOK actually cost per month?

For a typical indie founder doing 30-90 minutes of AI work per day, BYOK costs $2-15 per month with Claude, OpenAI, or Gemini. That's lower than the $20/month flat for ChatGPT Plus, Claude Pro, or Notion AI. Heavy users (multi-hour daily, large context) pay more, $20-40/month, but still typically less than stacked subscriptions.

Is BYOK more private than ChatGPT Plus?

Yes, in two specific ways. First, AI providers' API tier (which is what BYOK uses) typically does not train on your data by default, while consumer tiers often do. Second, with a BYOK desktop app, the app's company never sees your prompts; the requests go directly from your machine to the AI provider.

Where do I get an API key?

Anthropic Claude: console.anthropic.com → Settings → API Keys. OpenAI: platform.openai.com → API Keys. Google Gemini: aistudio.google.com → Get API key. Each takes 2-3 minutes to sign up and generate. All three require entering payment info to start, but the free tiers and prepaid balances are real (you can put $5 in and use it down).

What if I exhaust my API key budget?

All three providers let you set a hard monthly spending cap. If you set a $20 cap, the API will stop responding when you hit it instead of charging more. This protects you from runaway costs while you learn how much you actually use.

Which tools support BYOK?

Most desktop AI workspaces designed for power users support BYOK: Projelli, Cursor (for code), Continue.dev (for code), several Obsidian AI plugins, Raycast AI. Most cloud workspaces (Notion AI, ChatGPT Plus, Claude.ai Pro, Reflect, Mem.ai) do not.

Can I use multiple providers from the same tool?

If the tool supports multiple providers, yes. Projelli ships with all three (Claude, OpenAI, Gemini) plus an Ollama integration for local models. You add keys for each, and pick which provider to use on a per-conversation basis. The advantage: if one provider is down or rate-limited, you switch in one click.

What about Ollama and local models?

Ollama and LM Studio let you run open-weight models entirely on your machine. No API key, no internet, fully offline. Quality is competitive for code and short-form tasks; it lags behind Claude / GPT / Gemini for long-form strategic work. Most founders use a hybrid: BYOK Claude for high-stakes work, local Ollama for volume.

What to read next

Try BYOK with Projelli

Free download. $49 Pro one-time. $99 Lifetime. Founder's tier is $29 for the first 100 buyers. Bring your own Claude / OpenAI / Gemini key.

Get Projelli