Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Word Vector

Word Vector

Dense numerical representations of words encoding semantic meaning and linguistic relationships.

Year: 2013Generality: 0.72
Back to Vocab

A word vector, also called a word embedding, is a dense, fixed-length numerical representation of a word in a continuous vector space. Unlike sparse representations such as one-hot encoding—where each word is an isolated binary flag with no relationship to any other—word vectors place semantically related words near each other in high-dimensional space. This geometric property means that words like "king" and "queen" cluster together, and arithmetic operations on vectors can capture analogies: the classic example being that vec("king") − vec("man") + vec("woman") ≈ vec("queen"). Typical embedding dimensions range from 50 to 1,000 values, compressing vocabulary-scale information into compact, learnable representations.

Word vectors are learned by training shallow neural networks or matrix factorization methods on large text corpora, using the distributional hypothesis—the idea that words appearing in similar contexts tend to have similar meanings. Word2Vec, introduced by Mikolov et al. in 2013, popularized two training objectives: Continuous Bag-of-Words (CBOW), which predicts a target word from surrounding context, and Skip-gram, which predicts context words from a target. GloVe (Global Vectors for Word Representation) takes a complementary approach, factorizing a global word co-occurrence matrix to produce embeddings that reflect corpus-wide statistical patterns. FastText extended these ideas by representing words as bags of character n-grams, enabling meaningful vectors for rare or morphologically complex words.

Word vectors became a cornerstone of NLP pipelines throughout the 2010s, dramatically improving performance on tasks such as sentiment analysis, named entity recognition, machine translation, and question answering. Pre-trained embeddings allowed practitioners to transfer linguistic knowledge learned from billions of tokens into downstream models trained on far smaller datasets—an early and influential form of transfer learning in NLP.

While static word vectors assign a single representation per word regardless of context, they laid the conceptual groundwork for contextual embeddings produced by models like ELMo, BERT, and GPT, where a word's representation shifts depending on its surrounding sentence. Despite being largely superseded by these transformer-based approaches for state-of-the-art tasks, word vectors remain widely used for their computational efficiency, interpretability, and strong performance in resource-constrained settings.

Related

Related

Embedding
Embedding

A dense vector representation that encodes semantic relationships between discrete items.

Generality: 0.88
Contextual Embedding
Contextual Embedding

Word representations that dynamically shift meaning based on surrounding context.

Generality: 0.75
Embedding Space
Embedding Space

A learned vector space where similar data points cluster geometrically close together.

Generality: 0.79
Vectorization
Vectorization

Converting raw data into numerical vectors so machine learning algorithms can process it.

Generality: 0.72
Vector Database
Vector Database

A database optimized for storing and searching high-dimensional vector embeddings.

Generality: 0.62
BOW (Bag of Words)
BOW (Bag of Words)

A text representation encoding documents as unordered token counts, ignoring word sequence.

Generality: 0.73