Add comprehensive model card for Guidance-Free Training (GFT) with metadata
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,3 +1,59 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: mit
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
pipeline_tag: text-to-image
|
| 4 |
+
library_name: diffusers
|
| 5 |
+
---
|
| 6 |
+
|
| 7 |
+
# Guidance-Free Training (GFT): Visual Generation Without Guidance
|
| 8 |
+
|
| 9 |
+
This repository contains checkpoints and code for **Guidance-Free Training (GFT)**, a novel approach for visual generative models presented in the paper [Visual Generation Without Guidance](https://huggingface.co/papers/2501.15420). GFT aims to eliminate the need for Classifier-Free Guidance (CFG) during sampling, effectively halving the computational cost of inference while matching or surpassing CFG's performance.
|
| 10 |
+
|
| 11 |
+
Unlike previous distillation-based approaches, GFT enables training directly from scratch and requires minimal modifications to existing codebases. It is a universal algorithm applicable across various visual generative models, including diffusion, autoregressive, and masked-prediction architectures.
|
| 12 |
+
|
| 13 |
+
**Paper:** [Visual Generation Without Guidance](https://huggingface.co/papers/2501.15420)
|
| 14 |
+
**GitHub Repository:** https://github.com/thu-ml/GFT
|
| 15 |
+
|
| 16 |
+
<p align="center">
|
| 17 |
+
<img src="https://github.com/thu-ml/GFT/raw/main/GFT.png" alt="GFT comparison" style="width:80%;">
|
| 18 |
+
</p>
|
| 19 |
+
<p align="center">
|
| 20 |
+
<b>Qualitative T2I comparison between vanilla conditional generation, GFT, and CFG on Stable Diffusion 1.5 with the prompt "Elegant crystal vase holding pink peonies, soft raindrops tracing paths down the window behind it".</b>
|
| 21 |
+
</p>
|
| 22 |
+
|
| 23 |
+
## Key Features
|
| 24 |
+
|
| 25 |
+
* **Highly Efficient**: GFT reduces sampling to a single model inference, effectively halving the computational cost compared to CFG.
|
| 26 |
+
* **Minimal Modifications**: It requires fewer than 10 lines of code changes to existing visual generative model codebases, inheriting most design choices and hyperparameters.
|
| 27 |
+
* **Universal Applicability**: GFT is highly versatile, working across diverse visual generative models such as diffusion, Flow, autoregressive, and masked-prediction architectures.
|
| 28 |
+
* **Training from Scratch**: Unlike distillation methods, GFT enables direct training of guidance-free models from scratch.
|
| 29 |
+
* **Performance Match**: Consistently achieves comparable or even lower FID scores with similar diversity-fidelity trade-offs compared to CFG baselines.
|
| 30 |
+
* **Flexible Sampling**: Allows adjustment of sampling temperature with only a single model.
|
| 31 |
+
|
| 32 |
+
<p align="center">
|
| 33 |
+
<img src="https://github.com/thu-ml/GFT/raw/main/temperature.png" alt="temperature control" style="width:80%;">
|
| 34 |
+
</p>
|
| 35 |
+
<p align="center">
|
| 36 |
+
<b>GFT allows us to adjust sampling temperature of visual generation, with only a single model.</b>
|
| 37 |
+
</p>
|
| 38 |
+
|
| 39 |
+
## Usage and Pretrained Checkpoints
|
| 40 |
+
|
| 41 |
+
The project provides training code and pretrained guidance-free checkpoints for various models. These include:
|
| 42 |
+
* DiT models
|
| 43 |
+
* Stable Diffusion 1.5 models (e.g., [SD1.5-GF-finetune](https://huggingface.co/aaa-ceku7/GFT/tree/main/SD1.5-GF-finetune))
|
| 44 |
+
* LlamaGen models
|
| 45 |
+
|
| 46 |
+
For detailed implementation, training instructions, and example usage, please refer to the respective directories within the [GitHub repository](https://github.com/thu-ml/GFT).
|
| 47 |
+
|
| 48 |
+
## Citation
|
| 49 |
+
|
| 50 |
+
If you find our project helpful, please consider citing:
|
| 51 |
+
|
| 52 |
+
```bibtex
|
| 53 |
+
@article{chen2025visual,
|
| 54 |
+
title={Visual Generation Without Guidance},
|
| 55 |
+
author={Chen, Huayu and Jiang, Kai and Zheng, Kaiwen and Chen, Jianfei and Su, Hang and Zhu, Jun},
|
| 56 |
+
journal={arXiv preprint arXiv:2501.15420},
|
| 57 |
+
year={2025}
|
| 58 |
+
}
|
| 59 |
+
```
|