Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -3,14 +3,17 @@ import spaces
|
|
3 |
## Load model directly
|
4 |
# Load model directly
|
5 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
|
|
6 |
|
7 |
-
|
8 |
-
|
|
|
9 |
|
|
|
|
|
|
|
10 |
|
11 |
|
12 |
-
print(API_KEY)
|
13 |
-
|
14 |
|
15 |
|
16 |
|
|
|
3 |
## Load model directly
|
4 |
# Load model directly
|
5 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
6 |
+
from huggingface_hub import HfApi
|
7 |
|
8 |
+
# Récupérer le jeton d'accès API depuis Hugging Face Secrets
|
9 |
+
api = HfApi()
|
10 |
+
secret = api.secrets.get("alex-abb/8B-on-GPU-ZERO/API_KEY")
|
11 |
|
12 |
+
# Utiliser le jeton d'accès API pour charger le modèle
|
13 |
+
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Meta-Llama-3-8B-Instruct", token=secret["value"])
|
14 |
+
model = AutoModelForCausalLM.from_pretrained("meta-llama/Meta-Llama-3-8B-Instruct", token=secret["value"])
|
15 |
|
16 |
|
|
|
|
|
17 |
|
18 |
|
19 |
|