Awesome
<img src="https://t3bench.com/static/images/logobig.png" align="center"> <p align="center"> 📃 <a href="https://arxiv.org/abs/2310.02977" target="_blank">Paper</a> • 🌐 <a href="https://t3bench.com" target="_blank">Project Page</a> </p>T<sup>3</sup>Bench: Benchmarking Current Progress in Text-to-3D Generation
T<sup>3</sup>Bench is the first comprehensive text-to-3D benchmark containing diverse text prompts of three increasing complexity levels that are specially designed for 3D generation (300 prompts in total). To assess both the subjective quality and the text alignment, we propose two automatic metrics based on multi-view images produced by the 3D contents. The quality metric combines multi-view text-image scores and regional convolution to detect quality and view inconsistency. The alignment metric uses multi-view captioning and Large Language Model (LLM) evaluation to measure text-3D consistency. Both metrics closely correlate with different dimensions of human judgments, providing a paradigm for efficiently evaluating text-to-3D models.
<img src="https://t3bench.com/static/images/pipeline_v2.png">🔥 Updates
[2023/10/24] We have released mesh results of all prompt sets and methods! Please check <a href="https://drive.google.com/file/d/127Pfy6WI8txJU1DjmdpkR2y8-u5DoPIK/view?usp=share_link">here</a> to download.
Evaluate on T<sup>3</sup>Bench
Environment Setup
We adopt the implementation of <a href="https://github.com/threestudio-project/threestudio">ThreeStudio</a> to test the current text-to-3D methods. Please first follow the instructions of ThreeStudio to setup the generation environment.
Then install the following packages used for evaluation:
pip install -r requirements.txt
Note that we use a slightly modified version of ThreeStudio to ensure efficient generation.
Evaluation
Run Text-to-3D and Extract Mesh
# YOUR_GROUP: Choose the prompt set to test, including [single, surr, multi]
# YOUR_METHOD: We now support latentnerf, magic3d, fantasia3d, dreamfusion, sjc, and prolificdreamer.
python run_t3.py --group YOUR_GROUP --gpu YOUR_GPU --method YOUR_METHOD
python run_mesh.py --group YOUR_GROUP --gpu YOUR_GPU --method YOUR_METHOD
Quality Evaluation
python run_eval_quality.py --group YOUR_GROUP --gpu YOUR_GPU --method YOUR_METHOD
Alignment Evaluation
# First get the 3D prompt of the text-to-3D result
python run_caption.py --group YOUR_GROUP --gpu YOUR_GPU --method YOUR_METHOD
# then run the LLM Evaluation
python run_eval_alignment.py --group YOUR_GROUP --gpu YOUR_GPU --method YOUR_METHOD
Citation
@misc{he2023t3bench,
title={T$^3$Bench: Benchmarking Current Progress in Text-to-3D Generation},
author={Yuze He and Yushi Bai and Matthieu Lin and Wang Zhao and Yubin Hu and Jenny Sheng and Ran Yi and Juanzi Li and Yong-Jin Liu},
year={2023},
eprint={2310.02977},
archivePrefix={arXiv},
primaryClass={cs.CV}
}
Acknowledgement
This project could not be possible without the open-source works from <a href="https://github.com/threestudio-project/threestudio">ThreeStudio</a>, <a href="https://github.com/crockwell/Cap3D">Cap3D</a>, <a href="https://github.com/ashawkey/stable-dreamfusion">Stable-DreamFusion</a>, <a href="https://github.com/THUDM/ImageReward">ImageReward</a>, <a href="https://github.com/salesforce/LAVIS">LAVIS</a>. We sincerely thank them all.