Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -67,7 +67,7 @@ def avaliable_providers():
|
|
67 |
response = requests.get(endpoint_url, headers=headers)
|
68 |
|
69 |
for provider in response.json()['items']:
|
70 |
-
if provider['status']
|
71 |
providers.append(provider['vendor'])
|
72 |
|
73 |
return providers
|
@@ -118,7 +118,7 @@ def update_regions(provider):
|
|
118 |
response = requests.get(endpoint_url, headers=headers)
|
119 |
|
120 |
for region in response.json()['items']:
|
121 |
-
if region['status']
|
122 |
avalialbe_regions.append(f"{region['region']}/{region['label']}")
|
123 |
|
124 |
return gr.Dropdown.update(
|
@@ -126,10 +126,40 @@ def update_regions(provider):
|
|
126 |
value=avalialbe_regions[0] if len(avalialbe_regions) > 0 else None
|
127 |
)
|
128 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
129 |
with gr.Blocks() as demo2:
|
130 |
gr.Markdown(
|
131 |
"""
|
132 |
-
|
133 |
""")
|
134 |
hf_token_input = gr.Textbox(
|
135 |
label="enter your Hugging Face Access Token"
|
|
|
67 |
response = requests.get(endpoint_url, headers=headers)
|
68 |
|
69 |
for provider in response.json()['items']:
|
70 |
+
if provider['status'] == 'avaliable':
|
71 |
providers.append(provider['vendor'])
|
72 |
|
73 |
return providers
|
|
|
118 |
response = requests.get(endpoint_url, headers=headers)
|
119 |
|
120 |
for region in response.json()['items']:
|
121 |
+
if region['status'] == 'avaliable':
|
122 |
avalialbe_regions.append(f"{region['region']}/{region['label']}")
|
123 |
|
124 |
return gr.Dropdown.update(
|
|
|
126 |
value=avalialbe_regions[0] if len(avalialbe_regions) > 0 else None
|
127 |
)
|
128 |
|
129 |
+
def update_compute_options(provider, regeion):
|
130 |
+
# CPU [medium] 路 2vCPU 4GB 路 Intel Ice Lake
|
131 |
+
# 'https://api.endpoints.huggingface.cloud/provider/aws/region/us-east-1/compute'
|
132 |
+
"""
|
133 |
+
"accelerator": "cpu",
|
134 |
+
"numAccelerators": 1,
|
135 |
+
"memoryGb": "2Gi",
|
136 |
+
"instanceType": "c6i",
|
137 |
+
"instanceSize": "small",
|
138 |
+
"architecture": "Intel Ice Lake",
|
139 |
+
"status": "available",
|
140 |
+
"pricePerHour": 0.06
|
141 |
+
|
142 |
+
"""
|
143 |
+
|
144 |
+
avalialbe_compute_options = []
|
145 |
+
|
146 |
+
headers = {
|
147 |
+
"Content-Type": "application/json",
|
148 |
+
}
|
149 |
+
endpoint_url = f"https://api.endpoints.huggingface.cloud/provider/{provider}/region/{region}/compute"
|
150 |
+
response = requests.get(endpoint_url, headers=headers)
|
151 |
+
|
152 |
+
for compute in response.json()['items']:
|
153 |
+
if compute['status'] == 'avaliable':
|
154 |
+
avalialbe_compute_options.append(
|
155 |
+
f"{compute['accelerator'].upper()} []"
|
156 |
+
)
|
157 |
+
|
158 |
+
|
159 |
with gr.Blocks() as demo2:
|
160 |
gr.Markdown(
|
161 |
"""
|
162 |
+
## Your own Stable Diffusion on Hugging Face 馃 Endpoint
|
163 |
""")
|
164 |
hf_token_input = gr.Textbox(
|
165 |
label="enter your Hugging Face Access Token"
|