Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Token

Token

The basic unit of text that language models read, process, and generate.

Year: 1990Generality: 720
Back to Vocab

A token is the atomic unit of text that a natural language processing (NLP) system operates on. Rather than processing raw character streams or entire sentences at once, models consume sequences of tokens — discrete chunks that may correspond to words, subwords, punctuation marks, or individual characters depending on the tokenization scheme in use. The process of converting raw text into tokens is called tokenization, and it sits at the very front of nearly every NLP pipeline, shaping everything that follows.

Modern systems most commonly use subword tokenization algorithms such as Byte Pair Encoding (BPE), WordPiece, or SentencePiece. These methods strike a balance between character-level and word-level representations: frequent words are kept intact as single tokens, while rare or unknown words are split into smaller, recognizable pieces. This approach allows a fixed vocabulary of tens of thousands of tokens to cover virtually any input text, including novel words, technical jargon, and multiple languages, without resorting to an "unknown" catch-all symbol.

Tokens are the currency of large language models (LLMs). A model's context window — the amount of text it can attend to at once — is measured in tokens, not words or characters. Pricing for commercial APIs is typically quoted per token. Generation speed, memory consumption, and attention complexity all scale directly with token count. Understanding what constitutes a token in a given system is therefore practically important: a single English word averages roughly 1.3 tokens under common schemes, while code, non-Latin scripts, or whitespace-heavy formatting can tokenize far less efficiently.

The choice of tokenization strategy has measurable downstream effects on model performance. A vocabulary that poorly represents a language or domain forces the model to reconstruct meaning from many small fragments, increasing sequence length and making learning harder. Conversely, a well-matched tokenizer compresses text efficiently, reduces computational cost, and gives the model cleaner, more semantically coherent units to learn from. As multilingual and multimodal models have grown in importance, tokenization design has become an increasingly active area of research in its own right.

Related

Related

Token Processing
Token Processing

Segmenting text into discrete units that serve as inputs for NLP models.

Generality: 720
Price Per Token
Price Per Token

The unit cost charged for each token processed by a language model API.

Generality: 293
Tokenmaxxing
Tokenmaxxing

Maximizing useful information density within a prompt's token budget for better LLM outputs.

Generality: 94
Thought Token
Thought Token

Special tokens that give language models explicit space to reason before answering.

Generality: 450
Thinking Tokens
Thinking Tokens

Hidden reasoning tokens consumed during inference for internal step-by-step reasoning invisible to users

Generality: 605
Next Token Prediction
Next Token Prediction

A training objective where models learn to predict the next token in a sequence.

Generality: 794