ABDALLALSWAITI commited on
Commit
ce07205
·
verified ·
1 Parent(s): 3c3192e

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +12 -1
README.md CHANGED
@@ -19,7 +19,7 @@ base_model:
19
 
20
  # FLUX.1-dev-ControlNet-Union-Pro-2.0 (fp8)
21
 
22
- This repository contains an unified ControlNet for FLUX.1-dev model released by [Shakker Labs](https://huggingface.co/Shakker-Labs). This version has been quantized to FP8 format for optimized inference performance. We provide an [online demo](https://huggingface.co/spaces/Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro-2.0).
23
 
24
  # FP8 Quantization
25
  This model has been quantized from the original BFloat16 format to FP8 format. The benefits include:
@@ -27,6 +27,8 @@ This model has been quantized from the original BFloat16 format to FP8 format. T
27
  - **Faster Inference**: Potential speed improvements, especially on hardware with FP8 support
28
  - **Minimal Quality Loss**: Carefully calibrated quantization process to preserve output quality
29
 
 
 
30
  # Keynotes
31
  In comparison with [Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro](https://huggingface.co/Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro),
32
  - Remove mode embedding, has smaller model size.
@@ -157,6 +159,15 @@ pipe.to("cuda")
157
 
158
  See `fp8_inference_example.py` for a complete example.
159
 
 
 
 
 
 
 
 
 
 
160
  # Resources
161
  - [InstantX/FLUX.1-dev-IP-Adapter](https://huggingface.co/InstantX/FLUX.1-dev-IP-Adapter)
162
  - [InstantX/FLUX.1-dev-Controlnet-Canny](https://huggingface.co/InstantX/FLUX.1-dev-Controlnet-Canny)
 
19
 
20
  # FLUX.1-dev-ControlNet-Union-Pro-2.0 (fp8)
21
 
22
+ This repository contains an unified ControlNet for FLUX.1-dev model released by [Shakker Labs](https://huggingface.co/Shakker-Labs). This is a direct quantization of the original model to FP8 format for optimized inference performance (not a fine-tuned version). We provide an [online demo](https://huggingface.co/spaces/Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro-2.0).
23
 
24
  # FP8 Quantization
25
  This model has been quantized from the original BFloat16 format to FP8 format. The benefits include:
 
27
  - **Faster Inference**: Potential speed improvements, especially on hardware with FP8 support
28
  - **Minimal Quality Loss**: Carefully calibrated quantization process to preserve output quality
29
 
30
+ Note: This is a direct quantization of [Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro-2.0](https://huggingface.co/Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro-2.0) and preserves all the functionality of the original model.
31
+
32
  # Keynotes
33
  In comparison with [Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro](https://huggingface.co/Shakker-Labs/FLUX.1-dev-ControlNet-Union-Pro),
34
  - Remove mode embedding, has smaller model size.
 
159
 
160
  See `fp8_inference_example.py` for a complete example.
161
 
162
+ # Pushing Model to Hugging Face Hub
163
+ To push your FP8 quantized model to the Hugging Face Hub, use the included script:
164
+
165
+ ```bash
166
+ python push_model_to_hub.py --repo_id "ABDALLALSWAITI/FLUX.1-dev-ControlNet-Union-Pro-2.0-fp8"
167
+ ```
168
+
169
+ You will need to have the `huggingface_hub` library installed and be logged in with your Hugging Face credentials.
170
+
171
  # Resources
172
  - [InstantX/FLUX.1-dev-IP-Adapter](https://huggingface.co/InstantX/FLUX.1-dev-IP-Adapter)
173
  - [InstantX/FLUX.1-dev-Controlnet-Canny](https://huggingface.co/InstantX/FLUX.1-dev-Controlnet-Canny)