Official repository for the paper Can Test-Time Scaling Improve World Foundation Models?
- [2025.07] Paper is accepted by COLM 2025!
- [2025.04] Test-time scaling code released!
- [2025.04] Project website is live!
- [2025.03] Paper released on arXiv!
WFM-TTS is the first test-time scaling framework for World Foundation Models (WFMs). Instead of retraining or enlarging models, WFM-TTS improves performance at inference time using smart generation strategies. It:
- Enables small models (e.g., 4B) to match or outperform large models (e.g., 12B)
- Works under the same compute budget
- Requires no weight updates or additional training
We introduce a modular and extensible evaluation toolkit to assess WFM performance across:
- ✔ 3D consistency
- ✔ Temporal consistency
- ✔ Spatial relationship awareness
- ✔ Perceptual quality
- ✔ Text-to-video alignment
This toolkit provides rigorous benchmarking for generated videos across physical and semantic fidelity.
WFM-TTS integrates multiple test-time strategies to boost performance:
- Rule-Based Rewards — Robust and extensible scoring mechanisms
- Efficient Tokenizer Decoder — 9,000x faster than diffusion decoder with consistent trends
- Probability-Based Top-K Pruning — Balances exploration and quality
- Beam Search Integration — Enhances diversity and reliability
- A 4B WFM + WFM-TTS achieves better or equal performance to a 12B model
- Human evaluations favor WFM-TTS-enhanced outputs over larger baselines
git clone https://github.com/Mia-Cong/WFM-TTS.git
cd WFM-TTS
Base model Cosmos runs only on Linux systems. It has been tested with Ubuntu 20.04, 22.04, and 24.04. Python 3.10.x and conda are required.
Run the following to set up the conda environment and install dependencies:
# Create the WFM-TTS conda environment
conda env create --file wfmtts.yaml
# Activate the environment
conda activate wfmtts
# Install Python dependencies
pip install -r requirements.txt
# Patch Transformer engine linking issues in conda environments
ln -sf $CONDA_PREFIX/lib/python3.10/site-packages/nvidia/*/include/* $CONDA_PREFIX/include/
ln -sf $CONDA_PREFIX/lib/python3.10/site-packages/nvidia/*/include/* $CONDA_PREFIX/include/python3.10
# Install Transformer engine
pip install transformer-engine[pytorch]==1.12.0
To test the environment setup:
CUDA_HOME=$CONDA_PREFIX PYTHONPATH=$(pwd) python scripts/test_environment.py
To reproduce WFM-TTS's test-time scaling results:
- Follow cosmos_inference_autoregressive_base to set up COSMOS 4B/12B models.
- Follow cosmos_inference_autoregressive_video2world for COSMOS 5B/13B models.
- Download the 900 autonomous driving test sequences we prepared using NuScenes dataset and Waymo dataset, and put it under
assets/autoregressive/
- Run any of the provided test-time scaling scripts located in
scripts/
, for example:
./scripts/cosmos4b_prob_beam.sh
If you find this work useful, please cite:
@inproceedings{cong2025wfm-tts,
title = {Can Test-Time Scaling Improve World Foundation Models?},
author = {Wenyan Cong and Hanqing Zhu and Peihao Wang and Bangya Liu and Dejia Xu and Kevin Wang and David Z. Pan and Yan Wang and Zhiwen Fan and Zhangyang Wang},
booktitle = {COLM},
year = {2025}
}
For more updates and demos, visit our website: https://scalingwfm.github.io