Run with ComfyUI-Flux2-INT8

Quantized to INT8 using the following command:

convert_to_quant -i ./Chroma1-HD-Flash.safetensors -o ./Chroma1-HD-Flash-int8tensormixed.safetensors --int8 --scaling_mode tensor --distillation_large --comfy_quant --save-quant-metadata --optimizer radam --calib_samples 8192 --num_iter 3000 --lr "7.1267000000029e-4" --top_p 0.2 --min_k 256 --max_k 1024 --lr_schedule plateau --lr_patience 5 --lr_factor 0.92 --lr_min "1e-9" --lr_cooldown 1 --lr_threshold 1e-8 --early-stop-lr 8e-9 --lr-shape-influence 1.5 --low-memory
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for kcmaster/Chroma1-HD-Flash-INT8

Finetuned
(2)
this model