SAR3D: Autoregressive 3D Object Generation and Understanding via Multi-scale 3D VQVAE

1S-Lab, Nanyang Technological University
2Shanghai Artificial Intelligence Laboratory

SAR3D is a framework for fast 3D generation (<1s) and
detailed understanding via autoregressive modeling.

Abstract


Autoregressive models have demonstrated remarkable success across various fields, from large language models (LLMs) to large multimodal models (LMMs) and 2D content generation, moving closer to artificial general intelligence (AGI). Despite these advances, applying autoregressive approaches to 3D object generation and understanding remains largely unexplored. This paper introduces Scale AutoRegressive 3D (SAR3D), a novel framework that leverages a multi-scale 3D vector-quantized variational autoencoder (VQVAE) to tokenize 3D objects for efficient autoregressive generation and detailed understanding. By predicting the next scale in a multi-scale latent representation instead of the next single token, SAR3D reduces generation time significantly, achieving fast 3D object generation in just 0.82 seconds on an A6000 GPU. Additionally, given the tokens enriched with hierarchical 3D-aware information, we finetune a pretrained LLM on them, enabling multimodal comprehension of 3D content. Our experiments show that SAR3D surpasses current 3D generation methods in both speed and quality and allows LLMs to interpret and caption 3D models comprehensively.

Video


Single Image to 3D


Text to 3D


3D Captioning

3D Captioning Image

Simultaneous 3D Generation and Understanding


Method

Overview of Multi-scale VQVAE. Given a 3D model, we leverage multi-view RGB-D (depth) renderings and Plücker embeddings as the input to our multi-view encoder 𝓔. The encoder predicts a continuous feature map that is then quantized by the multi-scale quantizer 𝓠, giving R = (r₁, r₂, ..., rₖ) of latent tri-plane features. Each code of different scales shares the same codebook. The triplane decoder then converts the quantized latent triplane features into the triplane representation through a plane-wise manner. The predicted triplane is multi-view supervised with the ground truth image, depth, and normal.


3D Captioning Image


Overview of 3D Generation and 3D Understanding. Given a 3D model, our 3D VQVAE encodes it into multi-scale discrete tokens for both 3D generation and understanding.

In (a) 3D Generation, text or a single image is encoded by CLIPT or DINOv2, and the encoded condition features are integrated into the decoder-only transformer via cross attention. The transformer then causally predicts each scale of the latent triplane.

In (b) 3D Understanding, truncated 3D tokens are first processed with an MLP projector. The large language model receives a multimodal sequence of text and 3D tokens and generates a detailed caption describing the input 3D model.

3D Captioning Image

BibTeX


replace me