What to paste
Use this page for prompts, code, transcripts, article drafts, documentation, emails, JSON, markdown, and chat messages before sending them to an AI model.
The AI Token Counter estimates how many tokens your prompt, text, or chat message will use across GPT, Claude, Gemini, and other language models. It also shows characters, words, token chips, and approximate input cost without uploading anything.
Loading tokenizer...
Use this page for prompts, code, transcripts, article drafts, documentation, emails, JSON, markdown, and chat messages before sending them to an AI model.
Watch tokens, context used, remaining context, and estimated cost. Leave room for the model response, tool calls, citations, or JSON output.
Every model family tokenizes text differently. Code, symbols, whitespace, emojis, and non-English text can shift counts more than normal prose.
| Content type | Typical tokens | Useful for |
|---|---|---|
| Short prompt | 50-300 | Quick generation, classification, rewrite tasks |
| Email or memo | 150-600 | Summaries, replies, editing, tone checks |
| One page of prose | 600-900 | Article drafts, study notes, page analysis |
| Code file | 800-3,000+ | Review, refactor, test generation, debugging |
| Transcript hour | 8,000-12,000+ | Meeting summaries, call analysis, action items |
This utility runs as static HTML, CSS, and vanilla JavaScript. Shared model metadata is bundled in a local browser data file, pricing-aware math is calculated locally, and controls work without sending data to any API.
Pricing changes frequently. Last updated: April 27, 2026. Verify against official vendor pricing pages before relying on estimates.
A token is a small chunk of text that an AI model reads, often a word fragment, punctuation mark, or short word. English text averages about four characters per token, but exact counts depend on the tokenizer.
This browser tool uses a local estimate based on text length plus model-specific adjustment factors. Use official provider tokenizers for final billing-sensitive counts.
No. The token counter runs entirely in your browser and does not call any model provider API. Your text remains on your device unless you copy it elsewhere.
Different model families use different vocabularies and encodings. Code, symbols, whitespace, and multilingual text can change tokenization dramatically.
Some tokenizers split CJK text differently from English. Modern multilingual encodings are better, but exact counts still vary by model.
cl100k_base is an older OpenAI tokenizer family. o200k_base is a newer OpenAI tokenizer family used by GPT-4o style models and the GPT entries in this data file.
Yes. Chat messages include role and boundary tokens, so a chat prompt is slightly longer than the visible text alone.