Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. LLM (Large Language Model)

LLM (Large Language Model)

Massive neural networks trained on text to understand and generate human language.

Year: 2018Generality: 905
Back to Vocab

A Large Language Model (LLM) is a deep learning system trained on enormous corpora of text data to perform a wide range of natural language tasks. Built on the transformer architecture, LLMs learn statistical relationships between words, phrases, and concepts by predicting tokens in sequence across billions or trillions of training examples. The result is a model that encodes rich representations of language, world knowledge, and reasoning patterns within its parameters — which typically number in the billions for modern systems.

LLMs work by processing input text as a sequence of tokens and using self-attention mechanisms to weigh relationships between all tokens simultaneously. During training, the model adjusts its parameters to minimize prediction error across massive datasets drawn from books, websites, code repositories, and other sources. After pretraining, models are often fine-tuned or aligned using techniques like reinforcement learning from human feedback (RLHF) to make outputs more accurate, helpful, and safe. Inference involves sampling from the model's probability distribution over possible next tokens, producing fluent, contextually appropriate text.

The practical capabilities of LLMs are remarkably broad: they can answer questions, summarize documents, translate languages, write and debug code, draft creative content, and engage in multi-turn dialogue. Performance scales predictably with model size, dataset size, and compute — a relationship formalized in scaling laws research — which has driven a sustained push toward ever-larger models. GPT-2 (2019), GPT-3 (2020), and subsequent systems like PaLM, LLaMA, and GPT-4 demonstrated successive leaps in capability that surprised even their creators.

LLMs have become foundational infrastructure for modern AI applications, powering products used by hundreds of millions of people. They also raise important questions around factual accuracy, bias amplification, intellectual property, and misuse. Understanding their capabilities and limitations — including tendencies to hallucinate plausible-sounding but false information — is essential for deploying them responsibly. As the dominant paradigm in NLP and an increasingly central component of multimodal AI systems, LLMs represent one of the most consequential developments in the history of artificial intelligence.

Related

Related

DLMs (Deep Language Models)
DLMs (Deep Language Models)

Deep neural networks trained to understand, generate, and translate human language.

Generality: 796
MLLMs (Multimodal Large Language Models)
MLLMs (Multimodal Large Language Models)

AI systems that understand and generate content across text, images, audio, and more.

Generality: 794
LLA (Large Language Agent)
LLA (Large Language Agent)

An autonomous AI system combining large language models with goal-directed task execution.

Generality: 511
LVLMs (Large Vision Language Models)
LVLMs (Large Vision Language Models)

Large AI models that jointly understand and reason over images and text.

Generality: 694
L2M (Large Memory Model)
L2M (Large Memory Model)

A decoder-only Transformer with addressable auxiliary memory enabling reasoning far beyond its attention window.

Generality: 189
LCMs (Large Concept Models)
LCMs (Large Concept Models)

Large-scale models that represent and reason over abstract, compositional concepts rather than raw tokens.

Generality: 381