Spaces:
Running
Running
Commit
·
b42cb0b
1
Parent(s):
53b28ed
refactor: update benchmark comparison details in app.py for clarity and consistency, enhancing the description of FLUX-juiced and its performance evaluation
Browse files
app.py
CHANGED
@@ -54,7 +54,7 @@ with gr.Blocks("ParityError/Interstellar", fill_width=True, css=custom_css) as d
|
|
54 |
"""
|
55 |
# 📊 InferBench
|
56 |
|
57 |
-
We ran a comprehensive benchmark comparing FLUX-juiced with the “FLUX.1 [dev]” endpoints offered by:
|
58 |
|
59 |
- Replicate: https://replicate.com/black-forest-labs/flux-dev
|
60 |
- Fal: https://fal.ai/models/fal-ai/flux/dev
|
@@ -63,14 +63,14 @@ with gr.Blocks("ParityError/Interstellar", fill_width=True, css=custom_css) as d
|
|
63 |
|
64 |
All of these inference providers offer FLUX.1 [dev] implementations but they don’t always communicate about the optimisation methods used in the background, and most endpoint have different response times and performance measure.
|
65 |
|
66 |
-
For comparison purposes we used the same generation
|
67 |
|
68 |
- 28 inference steps
|
69 |
- 1024×1024 resolution
|
70 |
- Guidance scale of 3.5
|
71 |
- H100 GPU (80GB)—only reported by Replicate
|
72 |
|
73 |
-
Although we did test with this configuration and hardware, the applied compression methods work with different config and hardware too!
|
74 |
|
75 |
> We published a full blog post on the [InferBench and FLUX-juiced](https://www.pruna.ai/blog/flux-juiced-the-fastest-image-generation-endpoint).
|
76 |
"""
|
|
|
54 |
"""
|
55 |
# 📊 InferBench
|
56 |
|
57 |
+
We ran a comprehensive benchmark comparing our very own FLUX-juiced with the “FLUX.1 [dev]” endpoints offered by:
|
58 |
|
59 |
- Replicate: https://replicate.com/black-forest-labs/flux-dev
|
60 |
- Fal: https://fal.ai/models/fal-ai/flux/dev
|
|
|
63 |
|
64 |
All of these inference providers offer FLUX.1 [dev] implementations but they don’t always communicate about the optimisation methods used in the background, and most endpoint have different response times and performance measure.
|
65 |
|
66 |
+
For comparison purposes we used the same generation set-up for all the providers.
|
67 |
|
68 |
- 28 inference steps
|
69 |
- 1024×1024 resolution
|
70 |
- Guidance scale of 3.5
|
71 |
- H100 GPU (80GB)—only reported by Replicate
|
72 |
|
73 |
+
Although we did test with this specific Pruna configuration and hardware, the applied compression methods work with different config and hardware too!
|
74 |
|
75 |
> We published a full blog post on the [InferBench and FLUX-juiced](https://www.pruna.ai/blog/flux-juiced-the-fastest-image-generation-endpoint).
|
76 |
"""
|