8000 Update diffusers-quantization.md to include the Space link by sayakpaul · Pull Request #2899 · huggingface/blog · GitHub
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

Update diffusers-quantization.md to include the Space link #2899

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Jun 9, 2025
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions diffusers-quantization.md
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ Before we dive into the technical details of how various quantization backends i

We created a setup where you can provide a prompt, and we generate results using both the original, high-precision model (e.g., Flux-dev in BF16) and several quantized versions (BnB 4-bit, BnB 8-bit). The generated images are then presented to you and your challenge is to identify which ones came from the quantized models.

Try it out here!
Try it out [here](https://huggingface.co/spaces/diffusers/flux-quant) or below!
<gradio-app src="https://diffusers-flux-quant.hf.space"></gradio-app>

Often, especially with 8-bit quantization, the differences are subtle and may not be noticeable without close inspection. More aggressive quantization like 4-bit or lower might be more noticeable, but the results can still be good, especially considering the massive memory savings. NF4 often gives the best trade-off though.
Expand Down Expand Up @@ -592,4 +592,4 @@ Here's a quick guide to choosing a quantization backend:

Quantization significantly lowers the barrier to entry for using large diffusion models. Experiment with these backends to find the best balance of memory, speed, and quality for your needs.

*Acknowledgements: Thanks to [Chunte](https://huggingface.co/Chunte) for providing the thumbnail for this post.*
*Acknowledgements: Thanks to [Chunte](https://huggingface.co/Chunte) for providing the thumbnail for this post.*
0