ngcanh commited on
Commit
ff8c629
·
verified ·
1 Parent(s): 217523a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -5
app.py CHANGED
@@ -21,8 +21,7 @@ from langchain_community.llms import HuggingFaceHub
21
  import pandas as pd
22
 
23
  # Configuración del modelo
24
- MODEL_NAME = "mistralai/Mixtral-8x7B-Instruct-v0.1"
25
- model_name = "google/gemma-2-2b"
26
  TOKEN=os.getenv('HF_TOKEN')
27
  subprocess.run(["huggingface-cli", "login", "--token", TOKEN, "--add-to-git-credential"])
28
  ######
@@ -33,15 +32,17 @@ os.environ["HF_TOKEN"] = st.secrets["HF_TOKEN"]
33
  # Initialize tokenizer
34
  @st.cache_resource
35
  def load_model():
36
- # MODEL_NAME= "lmsys/vicuna-7b-v1.5"
 
 
37
  MODEL_NAME = "google/gemma-2b-it"
38
 
39
  model = AutoModelForCausalLM.from_pretrained(
40
- MODEL_NAME,\
41
  # quantization_config=nf4_config, # add config
42
  # torch_dtype=torch.bfloat16, # save memory using float16
43
  # low_cpu_mem_usage=True,
44
- token= TOKEN
45
  ).to("cuda")
46
 
47
  tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
 
21
  import pandas as pd
22
 
23
  # Configuración del modelo
24
+
 
25
  TOKEN=os.getenv('HF_TOKEN')
26
  subprocess.run(["huggingface-cli", "login", "--token", TOKEN, "--add-to-git-credential"])
27
  ######
 
32
  # Initialize tokenizer
33
  @st.cache_resource
34
  def load_model():
35
+ TOKEN=os.getenv('HF_TOKEN')
36
+ subprocess.run(["huggingface-cli", "login", "--token", TOKEN, "--add-to-git-credential"])
37
+ os.environ["HF_TOKEN"] = st.secrets["HF_TOKEN"]
38
  MODEL_NAME = "google/gemma-2b-it"
39
 
40
  model = AutoModelForCausalLM.from_pretrained(
41
+ MODEL_NAME
42
  # quantization_config=nf4_config, # add config
43
  # torch_dtype=torch.bfloat16, # save memory using float16
44
  # low_cpu_mem_usage=True,
45
+ # token= TOKEN
46
  ).to("cuda")
47
 
48
  tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)