Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Mirage Effect

Mirage Effect

When multimodal AI models produce confident visual analysis from images that were never provided

Year: 2026Generality: 542
Back to Vocab

The mirage effect is a failure mode in multimodal AI models where the system constructs detailed visual descriptions, diagnoses, and reasoning traces from images that were silently removed from the input. Unlike hallucination — which involves generating incorrect details about a real input — the mirage effect involves building an entire fabricated perceptual reality and reasoning from it with high confidence. The term was introduced in a 2026 Stanford paper (MIRAGE) co-authored by Fei-Fei Li.

The researchers tested frontier models including GPT-5.1, Gemini 3 Pro, and Claude Opus 4.5 across six major vision benchmarks, both medical and general. When all images were silently removed without changing prompts, the models continued to score 70-80% accuracy — describing nonexistent X-rays in detail, identifying fake nodules, and diagnosing conditions from text patterns alone. The models did not detect the absence of visual input. More troublingly, a text-only 3-billion-parameter model fine-tuned on the same benchmarks without any images outperformed all frontier multimodal models and even human radiologists on a held-out test set.

The findings expose a structural problem in how multimodal AI capabilities are evaluated: up to 77% of questions in standard vision benchmarks could be answered through text-pattern recognition alone, without any genuine visual understanding. This means that leaderboard scores, benchmark breakthroughs, and claims of multimodal capability may reflect linguistic shortcuts rather than actual perception. The practical implications are severe for high-stakes domains like medical imaging, where mirage-mode diagnoses showed systematic bias toward the most dangerous conditions — STEMI, melanoma, carcinoma — based on textual priors rather than visual evidence.

The mirage effect raises fundamental questions about what it means for a model to 'see.' When a system performs better in mirage mode — not knowing it is blind — than when explicitly told no image is present, the boundary between perception and confabulation becomes disturbingly unclear. The paper's benchmark cleanup methodology (B-Clean) and code are open-sourced, offering a path toward benchmarks that actually test visual understanding rather than text-pattern matching.

Related

Related

Hallucination
Hallucination

When AI models confidently generate plausible but factually incorrect or fabricated outputs.

Generality: 794
Model Collapse (Silent Collapse)
Model Collapse (Silent Collapse)

Progressive AI degradation caused by recursive training on AI-generated synthetic data.

Generality: 339
Multimodal
Multimodal

AI systems that process and integrate multiple data types like text, images, and audio.

Generality: 796
Generator-Verifier Gap
Generator-Verifier Gap

The asymmetry between an AI model's ability to generate versus verify outputs.

Generality: 416
Reasoning Instability
Reasoning Instability

When AI models produce inconsistent or contradictory reasoning across similar inputs.

Generality: 395
AI Effect
AI Effect

Achieved AI tasks are dismissed as 'not real intelligence,' perpetually moving the goalposts.

Generality: 520