CCockrum commited on
Commit
edd5165
Β·
verified Β·
1 Parent(s): 30d38df

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -34,7 +34,7 @@ if "follow_up" not in st.session_state:
34
  st.session_state.follow_up = ""
35
 
36
  # βœ… Initialize Hugging Face Model (Explicitly Set to CPU/GPU)
37
- def get_llm_hf_inference(model_id="meta-llama/Llama-2-7b-chat-hf", max_new_tokens=800, temperature=0.8):
38
  #mistralai/Mistral-7B-Instruct-v0.3
39
  return HuggingFaceEndpoint(
40
  repo_id=model_id,
@@ -91,7 +91,7 @@ def generate_follow_up(user_text):
91
  "Ensure it's concise and structured exactly as requested without extra commentary."
92
  )
93
 
94
- hf = get_llm_hf_inference(max_new_tokens=30, temperature=0.8) # πŸ”₯ Lower temp for consistency
95
  output = hf.invoke(input=prompt_text).strip()
96
 
97
  # βœ… Extract the relevant part using regex to remove unwanted symbols or truncations
@@ -120,7 +120,7 @@ def get_response(system_message, chat_history, user_text, max_new_tokens=800):
120
  return response, follow_up, chat_history, nasa_url
121
 
122
  # βœ… Invoke Hugging Face Model
123
- hf = get_llm_hf_inference(max_new_tokens=max_new_tokens, temperature=0.9)
124
 
125
  filtered_history = "\n".join(f"{msg['role']}: {msg['content']}" for msg in chat_history)
126
 
 
34
  st.session_state.follow_up = ""
35
 
36
  # βœ… Initialize Hugging Face Model (Explicitly Set to CPU/GPU)
37
+ def get_llm_hf_inference(model_id="meta-llama/Llama-2-7b-chat-hf", max_new_tokens=800, temperature=0.7):
38
  #mistralai/Mistral-7B-Instruct-v0.3
39
  return HuggingFaceEndpoint(
40
  repo_id=model_id,
 
91
  "Ensure it's concise and structured exactly as requested without extra commentary."
92
  )
93
 
94
+ hf = get_llm_hf_inference(max_new_tokens=30, temperature=0.3) # πŸ”₯ Lower temp for consistency
95
  output = hf.invoke(input=prompt_text).strip()
96
 
97
  # βœ… Extract the relevant part using regex to remove unwanted symbols or truncations
 
120
  return response, follow_up, chat_history, nasa_url
121
 
122
  # βœ… Invoke Hugging Face Model
123
+ hf = get_llm_hf_inference(max_new_tokens=max_new_tokens, temperature=0.7)
124
 
125
  filtered_history = "\n".join(f"{msg['role']}: {msg['content']}" for msg in chat_history)
126