CCockrum commited on
Commit
68a27ab
Β·
verified Β·
1 Parent(s): 8dc255c

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -34,7 +34,7 @@ if "follow_up" not in st.session_state:
34
  st.session_state.follow_up = ""
35
 
36
  # βœ… Initialize Hugging Face Model (Explicitly Set to CPU/GPU)
37
- def get_llm_hf_inference(model_id="meta-llama/Llama-2-7b-chat-hf", max_new_tokens=512, temperature=0.7):
38
  #mistralai/Mistral-7B-Instruct-v0.3
39
  return HuggingFaceEndpoint(
40
  repo_id=model_id,
@@ -91,7 +91,7 @@ def generate_follow_up(user_text):
91
  "Ensure it's concise and structured exactly as requested without extra commentary."
92
  )
93
 
94
- hf = get_llm_hf_inference(max_new_tokens=30, temperature=0.6) # πŸ”₯ Lower temp for consistency
95
  output = hf.invoke(input=prompt_text).strip()
96
 
97
  # βœ… Extract the relevant part using regex to remove unwanted symbols or truncations
@@ -104,7 +104,7 @@ def generate_follow_up(user_text):
104
  return cleaned_output
105
 
106
  # βœ… Main Response Function
107
- def get_response(system_message, chat_history, user_text, max_new_tokens=512):
108
  action = predict_action(user_text)
109
 
110
  # βœ… Handle NASA-Specific Queries
 
34
  st.session_state.follow_up = ""
35
 
36
  # βœ… Initialize Hugging Face Model (Explicitly Set to CPU/GPU)
37
+ def get_llm_hf_inference(model_id="meta-llama/Llama-2-7b-chat-hf", max_new_tokens=800, temperature=0.8):
38
  #mistralai/Mistral-7B-Instruct-v0.3
39
  return HuggingFaceEndpoint(
40
  repo_id=model_id,
 
91
  "Ensure it's concise and structured exactly as requested without extra commentary."
92
  )
93
 
94
+ hf = get_llm_hf_inference(max_new_tokens=30, temperature=0.8) # πŸ”₯ Lower temp for consistency
95
  output = hf.invoke(input=prompt_text).strip()
96
 
97
  # βœ… Extract the relevant part using regex to remove unwanted symbols or truncations
 
104
  return cleaned_output
105
 
106
  # βœ… Main Response Function
107
+ def get_response(system_message, chat_history, user_text, max_new_tokens=800):
108
  action = predict_action(user_text)
109
 
110
  # βœ… Handle NASA-Specific Queries