<RETURN_TO_BASE

Tencent Unveils PrimitiveAnything: Transforming 3D Shape Reconstruction with Auto-Regressive Primitive Generation

Tencent introduces PrimitiveAnything, a novel AI framework that reconstructs 3D shapes by sequentially generating geometric primitives, enhancing semantic understanding and generalization across diverse objects.

Understanding Shape Primitive Abstraction

Shape primitive abstraction simplifies complex 3D forms into basic geometric units, crucial for human visual perception and significant in computer vision and graphics. Existing 3D generation methods, including meshes, point clouds, and neural fields, produce high-fidelity models but often lack semantic depth and interpretability necessary for applications like robotics and scene understanding.

Challenges in Traditional Methods

Traditional approaches rely on optimization-based techniques that fit geometric primitives but tend to over-segment shapes semantically, or learning-based methods trained on limited, category-specific datasets that struggle to generalize. Early methods used simple shapes such as cuboids and cylinders, evolving into more expressive primitives like superquadrics. Yet, aligning abstraction methods with human cognition while maintaining broad applicability remains challenging.

Introducing PrimitiveAnything

Inspired by advances in 3D content generation using large datasets and auto-regressive transformers, Tencent researchers developed PrimitiveAnything. This novel framework treats shape abstraction as a generative task, sequentially assembling primitives to mimic human reasoning. It employs a decoder-only transformer conditioned on shape features to generate variable-length sequences of primitives.

Technical Innovations

PrimitiveAnything uses a unified, ambiguity-free parameterization for multiple primitive types, ensuring high geometric accuracy and efficient learning. The model encodes each primitive's type, position, rotation, and scale into discrete tokens, which a transformer predicts autoregressively. The cascaded decoder models dependencies between attributes, producing coherent primitive assemblies. Training optimizes cross-entropy losses, Chamfer Distance for reconstruction, and applies Gumbel-Softmax for differentiable sampling. Generation continues until an end-of-sequence token is reached, enabling flexible, human-like decomposition.

Dataset and Evaluation

The researchers created HumanPrim, a large dataset with 120,000 3D samples annotated with primitive assemblies. Evaluations using Chamfer Distance, Earth Mover’s Distance, Hausdorff Distance, Voxel-IoU, and segmentation metrics demonstrate that PrimitiveAnything outperforms existing optimization and learning-based methods. Ablation studies highlight the significance of each component. The framework also supports 3D content generation from text or images, offering user-friendly editing, high-quality modeling, and over 95% storage savings.

Potential Applications

PrimitiveAnything's efficient and modular design suits interactive 3D applications such as gaming, where performance and ease of manipulation are critical. Its ability to generalize across object categories and align with human abstraction patterns makes it a promising tool for robotics, scene understanding, and creative content generation.

For more details, explore the Paper, Demo, and GitHub page linked by the researchers. Follow their updates on Twitter and join the ML community discussions.

🇷🇺

Сменить язык

Читать эту статью на русском

Переключить на Русский