Update app.py
Browse files
app.py
CHANGED
@@ -7,7 +7,7 @@ def get_hf_model_args(hf_model_name_or_path):
|
|
7 |
try:
|
8 |
config = AutoConfig.from_pretrained(hf_model_name_or_path, trust_remote_code=True).to_dict()
|
9 |
except Exception as e:
|
10 |
-
|
11 |
|
12 |
# Extract relevant values from the config
|
13 |
num_layers = config.get("num_hidden_layers", None)
|
@@ -22,14 +22,11 @@ def get_hf_model_args(hf_model_name_or_path):
|
|
22 |
"num_attention_heads": num_attention_heads,
|
23 |
"vocab_size": vocab_size,
|
24 |
"sequence_length": sequence_length,
|
25 |
-
}
|
26 |
|
27 |
# ---- Update Gradio inputs with Hugging Face model config ---- #
|
28 |
def update_from_hf_model(hf_model_name_or_path):
|
29 |
-
model_params
|
30 |
-
|
31 |
-
if hf_error:
|
32 |
-
return gr.update(), gr.update(), gr.update(), gr.update(), gr.update(), hf_error
|
33 |
|
34 |
return (gr.update(value=model_params["num_layers"]),
|
35 |
gr.update(value=model_params["hidden_size"]),
|
@@ -40,10 +37,7 @@ def update_from_hf_model(hf_model_name_or_path):
|
|
40 |
|
41 |
# ---- Memory Calculation ---- #
|
42 |
def calc_mem(hf_model_name_or_path, num_gpus, tensor_parallel_size, pipeline_parallel_size, batch_size_per_gpu, sequence_length, vocab_size, hidden_size, num_attention_heads, num_layers, ffn_expansion_factor, is_mixed_precision, misc_mem_gib):
|
43 |
-
model_params
|
44 |
-
|
45 |
-
if hf_error:
|
46 |
-
return hf_error
|
47 |
|
48 |
if model_params:
|
49 |
num_layers = model_params["num_layers"] or num_layers
|
|
|
7 |
try:
|
8 |
config = AutoConfig.from_pretrained(hf_model_name_or_path, trust_remote_code=True).to_dict()
|
9 |
except Exception as e:
|
10 |
+
raise gr.Error(f"Error fetching Hugging Face model: {str(e)}")
|
11 |
|
12 |
# Extract relevant values from the config
|
13 |
num_layers = config.get("num_hidden_layers", None)
|
|
|
22 |
"num_attention_heads": num_attention_heads,
|
23 |
"vocab_size": vocab_size,
|
24 |
"sequence_length": sequence_length,
|
25 |
+
}
|
26 |
|
27 |
# ---- Update Gradio inputs with Hugging Face model config ---- #
|
28 |
def update_from_hf_model(hf_model_name_or_path):
|
29 |
+
model_params = get_hf_model_args(hf_model_name_or_path)
|
|
|
|
|
|
|
30 |
|
31 |
return (gr.update(value=model_params["num_layers"]),
|
32 |
gr.update(value=model_params["hidden_size"]),
|
|
|
37 |
|
38 |
# ---- Memory Calculation ---- #
|
39 |
def calc_mem(hf_model_name_or_path, num_gpus, tensor_parallel_size, pipeline_parallel_size, batch_size_per_gpu, sequence_length, vocab_size, hidden_size, num_attention_heads, num_layers, ffn_expansion_factor, is_mixed_precision, misc_mem_gib):
|
40 |
+
model_params = get_hf_model_args(hf_model_name_or_path) if hf_model_name_or_path else None
|
|
|
|
|
|
|
41 |
|
42 |
if model_params:
|
43 |
num_layers = model_params["num_layers"] or num_layers
|