A model that transforms three-dimensional input data into a new 3D output.
A 3D-to-3D model is a machine learning architecture designed to accept three-dimensional data as input and produce transformed or synthesized three-dimensional data as output. Unlike models that map between fundamentally different modalities—such as images to text—3D-to-3D models operate entirely within the spatial domain, making them essential for tasks like shape completion, mesh deformation, point cloud upsampling, scene reconstruction, and volumetric segmentation. The input and output representations may share the same format, such as point cloud to point cloud, or differ, such as a voxel grid transformed into a mesh.
These models typically rely on neural network architectures adapted for the irregular and high-dimensional nature of 3D data. PointNet and its successors process unordered point clouds by learning per-point features and aggregating them globally. Graph neural networks treat mesh vertices and edges as nodes and connections, enabling message-passing over surface geometry. Volumetric CNNs apply convolutions over 3D grids but are computationally expensive, while implicit representations like Neural Radiance Fields (NeRF) and occupancy networks encode geometry as continuous functions, allowing high-fidelity output at arbitrary resolution. Diffusion models and transformer-based architectures have more recently been applied to 3D generation and editing tasks with strong results.
The practical importance of 3D-to-3D models spans numerous fields. In medical imaging, they enable organ segmentation and anatomical reconstruction from CT or MRI scans. In autonomous driving and robotics, they support real-time scene understanding and object manipulation from LiDAR point clouds. In computer graphics and digital content creation, they power tools for shape stylization, texture synthesis, and geometry refinement. As 3D capture hardware—including depth cameras, LiDAR sensors, and photogrammetry pipelines—becomes more accessible, the volume of raw 3D data available for training and deployment has grown substantially, accelerating progress in this area.
The field gained significant momentum in the mid-2010s with the introduction of deep learning methods capable of handling 3D structure, and has matured rapidly since 2020 with the rise of generative models and large-scale 3D datasets. Today, 3D-to-3D modeling sits at the intersection of geometry processing, computer vision, and generative AI, representing one of the more technically demanding frontiers in applied deep learning.