Skip to content

Conversation

kohya-ss
Copy link
Owner

@kohya-ss kohya-ss commented Sep 25, 2025

Add --fp8_scaled (block-wise fp8 quantization) for FLUX.1 and Chroma models.

See kohya-ss/musubi-tuner#564 for details about fp8 quantization.

Use --fp8_scaled instead of --fp8_base for flux_train_network.py.

@kohya-ss
Copy link
Owner Author

Comparison of inference results with Chroma1-HD.

bfloat16:
bf16_20250926_083953

fp8_scaled (block-wise quantization):
fp8_scaled_20250926_084213

fp8 (no quantization):
fp8_20250926_084057

The generation settings are the same as here, but the noise is generated on CUDA: https://huggingface.co/lodestones/Chroma1-HD#how-to-use

@iqddd
Copy link

iqddd commented Oct 6, 2025

But why use_scaled_mm=False?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants