File size: 1,562 Bytes
8bef849
 
ee70744
8bef849
 
 
47e69e1
 
 
 
 
 
 
 
8bef849
 
 
 
 
 
34c7e1a
8bef849
 
 
 
 
a8a64d5
8bef849
 
 
 
 
 
 
 
 
 
 
 
 
ee70744
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
---
license: apache-2.0
pipeline_tag: image-to-3d
tags:
- text-to-3d
- image-to-3d
datasets:
- fka/awesome-chatgpt-prompts
language:
- aa
metrics:
- accuracy
base_model: microsoft/Phi-3.5-MoE-instruct
library_name: diffusers
---

# 3DTopia-XL

This repo contains the pretrained weights for *3DTopia-XL: Scaling High-quality 3D Asset Generation via Primitive Diffusion*.

[Project Page]() | [Arxiv](https://arxiv.org/abs/xxxx.xxxxx) | [Weights](https://huggingface.co/FrozenBurning/3DTopia-XL) | [Code](https://github.com/3DTopia/3DTopia-XL)


## Introduction
3DTopia-XL scales high-quality 3D asset generation using Diffusion Transformer (DiT) built upon an expressive and efficient 3D representation, **PrimX**. The denoising process takes 5 seconds to generate a 3D PBR asset from text/image input which is ready for the graphics pipeline to use.

<video controls autoplay src="https://cdn-uploads.huggingface.co/production/uploads/62fc8cf7ee999004b5a8b982/-f349zLT7hjWla9yxupSY.mp4"></video>

## Model Details
The model is trained on a ~256K subset of [Objaverse](https://huggingface.co/datasets/allenai/objaverse).
For more details, please refer to our paper.

## Usage

To download the model:
```python
from huggingface_hub import hf_hub_download
ckpt_path = hf_hub_download(repo_id="frozenburning/3DTopia-XL", filename="model_sview_dit_fp16.pt")
vae_ckpt_path = hf_hub_download(repo_id="frozenburning/3DTopia-XL", filename="model_vae_fp16.pt")
```
Please refer to our [repo](https://github.com/3DTopia/3DTopia-XL) for more details on loading and inference.