Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Research
  3. Liminal
  4. Embodied AI Agents

Embodied AI Agents

AI systems that perceive and navigate 3D spaces like physical or virtual worlds
Back to LiminalView interactive version

Embodied AI agents represent a significant evolution in artificial intelligence, moving beyond text-based chatbots and voice assistants to create autonomous virtual entities capable of perceiving, reasoning about, and acting within three-dimensional environments. Unlike traditional AI systems that process information in abstract digital spaces, these agents possess spatial awareness and can navigate physical or virtual worlds with an understanding of geometry, object relationships, and environmental constraints. The technology combines advances in computer vision, spatial mapping, natural language processing, and reinforcement learning to create AI entities that can perceive their surroundings through sensors or virtual cameras, build mental models of 3D spaces, and execute physical or simulated actions. These agents employ sophisticated algorithms for path planning, obstacle avoidance, and task execution, while maintaining persistent memory of their environment and interactions. The embodiment aspect is crucial—these AI systems are designed to occupy and interact with space in ways that mirror physical presence, whether manifested as holographic figures in augmented reality, avatars in virtual environments, or digital representations overlaid on real-world spaces through mixed reality displays.

The emergence of embodied AI agents addresses several critical challenges in human-computer interaction and spatial computing adoption. Traditional interfaces for navigating complex 3D environments or managing smart spaces often require users to master unintuitive controls or abstract menu systems. Embodied agents solve this by providing natural, conversational interfaces that can guide users through physical locations, demonstrate procedures through gesture and movement, and serve as persistent assistants that understand spatial context. In enterprise settings, these agents enable more intuitive training simulations, where virtual instructors can demonstrate equipment operation or safety procedures in realistic spatial contexts. For consumer applications, they transform how people interact with smart homes and mixed reality entertainment, offering companions or guides that can move through spaces, point out features, and respond to both verbal commands and physical gestures. This technology also enables new forms of remote collaboration, where AI agents can represent absent team members or provide intelligent facilitation in virtual meeting spaces, understanding social dynamics and spatial arrangements.

Early deployments of embodied AI agents are emerging across multiple sectors, from retail environments where virtual assistants help customers navigate stores and locate products, to healthcare settings where AI companions provide guidance to patients in rehabilitation exercises. Research institutions are developing increasingly sophisticated agents capable of complex spatial reasoning, with some systems demonstrating the ability to learn new environments through exploration and adapt their behaviour based on user preferences and interaction patterns. The technology is particularly promising for accessibility applications, where embodied agents can serve as persistent guides for individuals with visual or cognitive impairments, providing spatial orientation and navigation assistance in both physical and virtual spaces. As mixed reality headsets become more prevalent and spatial computing platforms mature, industry observers note a trajectory toward agents that can seamlessly transition between purely virtual existence and augmented reality manifestations, maintaining continuity of relationship and knowledge across different spatial contexts. This evolution suggests a future where AI assistance becomes fundamentally spatial rather than screen-bound, with virtual entities that inhabit our environments as persistent, helpful presences rather than tools we must explicitly invoke.

TRL
3/9Conceptual
Impact
4/5
Investment
4/5
Category
Software

Related Organizations

Allen Institute for AI (AI2) logo
Allen Institute for AI (AI2)

United States · Nonprofit

95%

Creator of Semantic Scholar and various open-source models for scientific text processing.

Researcher
Google DeepMind logo
Google DeepMind

United Kingdom · Research Lab

95%

Developers of the Gemini family of models, which are trained from the start to be multimodal across text, images, video, and audio.

Developer
Inworld AI logo
Inworld AI

United States · Startup

90%

A platform for creating AI characters with distinct personalities, memories, and contextual awareness for games and virtual worlds.

Developer
Meta logo
Meta

United States · Company

90%

Developer of the Llama series of open-source LLMs.

Developer
Stanford University (HAI) logo
Stanford University (HAI)

United States · University

90%

Human-Centered AI Institute conducting research on BEHAVIOR benchmark.

Researcher
The AI Institute logo
The AI Institute

United States · Research Lab

90%

A research organization founded by Boston Dynamics, focusing on solving the most difficult problems in robotics and embodied AI.

Researcher
Convai logo
Convai

United States · Startup

85%

Provides conversational AI for virtual worlds, enabling NPCs to have voice-based interactions with players.

Developer
Covariant logo
Covariant

United States · Startup

85%

AI robotics company building a universal AI brain for robots.

Developer
Sanctuary AI logo
Sanctuary AI

Canada · Startup

85%

Developing general-purpose humanoid robots (Phoenix) powered by Carbon, their AI control system.

Developer
Embodied, Inc. logo
Embodied, Inc.

United States · Company

80%

Creators of Moxie, a companion robot that uses machine learning to perceive, process, and respond to natural conversation and eye contact to help children with social development.

Developer

Supporting Evidence

Evidence data is not available for this technology yet.

Connections

Software
Software
Avatar Embodiment Systems

Real-time systems translating human motion and expression into digital avatars

TRL
4/9
Impact
4/5
Investment
3/5
Software
Software
Spatial Foundation Models

AI models trained on 3D environments to understand spatial relationships and physical interactions

TRL
3/9
Impact
5/5
Investment
5/5
Software
Software
Semantic Scene Understanding

Real-time spatial comprehension of rooms, objects, and their functional relationships

TRL
6/9
Impact
5/5
Investment
5/5
Applications
Applications
Assistive Spatial Navigation

XR systems that guide blind, low-vision, and mobility-impaired users through physical spaces

TRL
6/9
Impact
5/5
Investment
3/5
Software
Software
Spatial Operating Systems

Operating systems that organize apps and data in 3D space instead of flat screens

TRL
6/9
Impact
5/5
Investment
5/5
Software
Software
Generative Physics Engines

Machine learning models that infer and adapt physical behaviors in virtual environments in real-time

TRL
2/9
Impact
3/5
Investment
3/5

Book a research session

Bring this signal into a focused decision sprint with analyst-led framing and synthesis.
Research Sessions