davidberenstein1957 commited on
Commit
b42cb0b
·
1 Parent(s): 53b28ed

refactor: update benchmark comparison details in app.py for clarity and consistency, enhancing the description of FLUX-juiced and its performance evaluation

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -54,7 +54,7 @@ with gr.Blocks("ParityError/Interstellar", fill_width=True, css=custom_css) as d
54
  """
55
  # 📊 InferBench
56
 
57
- We ran a comprehensive benchmark comparing FLUX-juiced with the “FLUX.1 [dev]” endpoints offered by:
58
 
59
  - Replicate: https://replicate.com/black-forest-labs/flux-dev
60
  - Fal: https://fal.ai/models/fal-ai/flux/dev
@@ -63,14 +63,14 @@ with gr.Blocks("ParityError/Interstellar", fill_width=True, css=custom_css) as d
63
 
64
  All of these inference providers offer FLUX.1 [dev] implementations but they don’t always communicate about the optimisation methods used in the background, and most endpoint have different response times and performance measure.
65
 
66
- For comparison purposes we used the same generation configuration and hardware among the different providers.
67
 
68
  - 28 inference steps
69
  - 1024×1024 resolution
70
  - Guidance scale of 3.5
71
  - H100 GPU (80GB)—only reported by Replicate
72
 
73
- Although we did test with this configuration and hardware, the applied compression methods work with different config and hardware too!
74
 
75
  > We published a full blog post on the [InferBench and FLUX-juiced](https://www.pruna.ai/blog/flux-juiced-the-fastest-image-generation-endpoint).
76
  """
 
54
  """
55
  # 📊 InferBench
56
 
57
+ We ran a comprehensive benchmark comparing our very own FLUX-juiced with the “FLUX.1 [dev]” endpoints offered by:
58
 
59
  - Replicate: https://replicate.com/black-forest-labs/flux-dev
60
  - Fal: https://fal.ai/models/fal-ai/flux/dev
 
63
 
64
  All of these inference providers offer FLUX.1 [dev] implementations but they don’t always communicate about the optimisation methods used in the background, and most endpoint have different response times and performance measure.
65
 
66
+ For comparison purposes we used the same generation set-up for all the providers.
67
 
68
  - 28 inference steps
69
  - 1024×1024 resolution
70
  - Guidance scale of 3.5
71
  - H100 GPU (80GB)—only reported by Replicate
72
 
73
+ Although we did test with this specific Pruna configuration and hardware, the applied compression methods work with different config and hardware too!
74
 
75
  > We published a full blog post on the [InferBench and FLUX-juiced](https://www.pruna.ai/blog/flux-juiced-the-fastest-image-generation-endpoint).
76
  """