Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
@@ -19,7 +19,7 @@ base_model:
|
|
19 |
|
20 |
# FLUX.1-dev-ControlNet-Union-Pro-2.0 (fp8)
|
21 |
|
22 |
-
This repository contains an unified ControlNet for FLUX.1-dev model released by [Shakker Labs](https://huggingface.co/Shakker-Labs). This
|
23 |
|
24 |
# FP8 Quantization
|
25 |
This model has been quantized from the original BFloat16 format to FP8 format. The benefits include:
|
@@ -27,6 +27,8 @@ This model has been quantized from the original BFloat16 format to FP8 format. T
|
|
27 |
- **Faster Inference**: Potential speed improvements, especially on hardware with FP8 support
|
28 |
- **Minimal Quality Loss**: Carefully calibrated quantization process to preserve output quality
|
29 |
|
|
|
|
|
30 |
# Keynotes
|
31 |
In comparison with [Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro](https://huggingface.co/Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro),
|
32 |
- Remove mode embedding, has smaller model size.
|
@@ -157,6 +159,15 @@ pipe.to("cuda")
|
|
157 |
|
158 |
See `fp8_inference_example.py` for a complete example.
|
159 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
160 |
# Resources
|
161 |
- [InstantX/FLUX.1-dev-IP-Adapter](https://huggingface.co/InstantX/FLUX.1-dev-IP-Adapter)
|
162 |
- [InstantX/FLUX.1-dev-Controlnet-Canny](https://huggingface.co/InstantX/FLUX.1-dev-Controlnet-Canny)
|
|
|
19 |
|
20 |
# FLUX.1-dev-ControlNet-Union-Pro-2.0 (fp8)
|
21 |
|
22 |
+
This repository contains an unified ControlNet for FLUX.1-dev model released by [Shakker Labs](https://huggingface.co/Shakker-Labs). This is a direct quantization of the original model to FP8 format for optimized inference performance (not a fine-tuned version). We provide an [online demo](https://huggingface.co/spaces/Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro-2.0).
|
23 |
|
24 |
# FP8 Quantization
|
25 |
This model has been quantized from the original BFloat16 format to FP8 format. The benefits include:
|
|
|
27 |
- **Faster Inference**: Potential speed improvements, especially on hardware with FP8 support
|
28 |
- **Minimal Quality Loss**: Carefully calibrated quantization process to preserve output quality
|
29 |
|
30 |
+
Note: This is a direct quantization of [Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro-2.0](https://huggingface.co/Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro-2.0) and preserves all the functionality of the original model.
|
31 |
+
|
32 |
# Keynotes
|
33 |
In comparison with [Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro](https://huggingface.co/Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro),
|
34 |
- Remove mode embedding, has smaller model size.
|
|
|
159 |
|
160 |
See `fp8_inference_example.py` for a complete example.
|
161 |
|
162 |
+
# Pushing Model to Hugging Face Hub
|
163 |
+
To push your FP8 quantized model to the Hugging Face Hub, use the included script:
|
164 |
+
|
165 |
+
```bash
|
166 |
+
python push_model_to_hub.py --repo_id "ABDALLALSWAITI/FLUX.1-dev-ControlNet-Union-Pro-2.0-fp8"
|
167 |
+
```
|
168 |
+
|
169 |
+
You will need to have the `huggingface_hub` library installed and be logged in with your Hugging Face credentials.
|
170 |
+
|
171 |
# Resources
|
172 |
- [InstantX/FLUX.1-dev-IP-Adapter](https://huggingface.co/InstantX/FLUX.1-dev-IP-Adapter)
|
173 |
- [InstantX/FLUX.1-dev-Controlnet-Canny](https://huggingface.co/InstantX/FLUX.1-dev-Controlnet-Canny)
|