The short answer: about 1.3 tokens per word in English. That holds across GPT, Claude, and Gemini for normal prose. The longer answer is more interesting — and matters if you are estimating costs, sizing context windows, or trimming prompts.
| English text | Approximate tokens |
|---|---|
| 1 word | 1.3 tokens |
| 10 words | 13 tokens |
| 100 words | 130 tokens |
| 500 words | 650 tokens |
| 1,000 words | 1,300 tokens |
| 1 page (~250 words) | 325 tokens |
| 5 pages (~1,250 words) | 1,625 tokens |
| 1 short book (~50,000 words) | 65,000 tokens |
Or going the other way:
| Tokens | Approximate words | Equivalent |
|---|---|---|
| 100 | 75 | One paragraph |
| 500 | 375 | Half a page |
| 1,000 | 750 | One full page |
| 4,000 | 3,000 | One short blog post |
| 8,000 | 6,000 | One long article |
| 32,000 | 24,000 | One short report |
| 128,000 | 96,000 | One short novel |
| 200,000 | 150,000 | One full novel |
Paste your text and see exact token count instantly.
Open Token Counter →Tokenizers do not split text on spaces. They split on a learned vocabulary of common subword pieces. This means:
| Content | Words per token | Why |
|---|---|---|
| Plain English prose | 0.75 | Standard tokenizer training data |
| Technical writing | 0.65 | More long words, jargon |
| Code (Python, JS) | 0.50 | Symbols and identifiers split heavily |
| Spanish/French | 0.55 | Different tokenizer coverage |
| Chinese/Japanese | 0.35 | Each character is often 1 token |
| Math/equations | 0.30 | Symbols are individual tokens |
| URLs and emails | 0.25 | Characters split heavily |
| JSON output | 0.45 | Brackets, commas, quotes all count |
If you're processing code or non-English text, your token count will be much higher than the English rule of thumb. For Chinese or Japanese, 1 word is often 2-3 tokens.
Each major LLM uses a slightly different tokenizer. The same 1,000-word English document tokenizes to different counts:
| Model | Tokenizer | Approx tokens (1,000 English words) |
|---|---|---|
| GPT-4o, GPT-4.1 | o200k_base | ~1,250 |
| GPT-3.5, older GPT-4 | cl100k_base | ~1,300 |
| Claude (all) | Claude tokenizer | ~1,290 |
| Gemini | SentencePiece | ~1,280 |
| Llama 3, 4 | BPE | ~1,310 |
| DeepSeek | BPE variant | ~1,295 |
The variation is small — within 5% — but it can matter at scale. If you are budgeting for 1 million prompts, a 5% difference is 50,000 tokens, which adds up.
The fastest way is to use a free online token counter:
The counter runs entirely in your browser — your text never leaves your device, so it works for confidential prompts and proprietary content.
Count tokens for any text in your browser. Free, no signup.
Open Token Counter →For most casual use, 1.3 tokens per word is fine. The difference matters when:
For everything else, the rule of thumb is fine. 1.3 tokens per word, 0.75 words per token. Use the counter when you need precision, use the rule when you need a rough estimate.