Niansuh commited on
Commit
e56ea0f
·
verified ·
1 Parent(s): 11a3d84

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -7
app.py CHANGED
@@ -7,7 +7,7 @@ import random
7
  app = Flask(__name__)
8
 
9
  def generate_random_ip():
10
- return f"{random.randint(1,255)}.{random.randint(0,255)}.{random.randint(0,255)}.{random.randint(0,255)}"
11
 
12
  def generate_user_agent():
13
  os_list = ['Windows NT 10.0', 'Windows NT 6.1', 'Mac OS X 10_15_7', 'Ubuntu', 'Linux x86_64']
@@ -52,7 +52,6 @@ def chat_completions():
52
  return {"error": "No messages provided"}, 400
53
 
54
  model = data.get('model', 'gpt-4o')
55
-
56
  if model.startswith('gpt'):
57
  endpoint = "openAI"
58
  original_api_url = 'https://chatpro.ai-pro.org/api/ask/openAI'
@@ -70,7 +69,6 @@ def chat_completions():
70
  }
71
 
72
  def generate():
73
- nonlocal messages
74
  full_response = ""
75
  while True:
76
  conversation = "\n".join([f"{msg['role']}: {msg['content']}" for msg in messages])
@@ -104,16 +102,14 @@ def chat_completions():
104
  messages.append({"role": "user", "content": "Please continue your output and do not repeat the previous content"})
105
  break # Jump out of the current loop and continue with the next request
106
  else:
107
- # End normally, sending the final content (if any)
108
  last_content = response_message.get('text', '')
109
  if last_content and last_content != full_response:
110
  yield f"data: {json.dumps(format_openai_response(last_content[len(full_response):]))}\n\n"
111
 
112
  yield f"data: {json.dumps(format_openai_response('', finish_reason))}\n\n"
113
  yield "data: [DONE]\n\n"
114
- return # completely end generation
115
 
116
- # If it ends due to multiple length limits, send a stop signal
117
  yield f"data: {json.dumps(format_openai_response('', 'stop'))}\n\n"
118
  yield "data: [DONE]\n\n"
119
 
@@ -153,4 +149,4 @@ def chat_completions():
153
  }
154
 
155
  if __name__ == '__main__':
156
- app.run(debug=True, port=5000)
 
7
  app = Flask(__name__)
8
 
9
  def generate_random_ip():
10
+ return f"{random.randint(1, 255)}.{random.randint(0, 255)}.{random.randint(0, 255)}.{random.randint(0, 255)}"
11
 
12
  def generate_user_agent():
13
  os_list = ['Windows NT 10.0', 'Windows NT 6.1', 'Mac OS X 10_15_7', 'Ubuntu', 'Linux x86_64']
 
52
  return {"error": "No messages provided"}, 400
53
 
54
  model = data.get('model', 'gpt-4o')
 
55
  if model.startswith('gpt'):
56
  endpoint = "openAI"
57
  original_api_url = 'https://chatpro.ai-pro.org/api/ask/openAI'
 
69
  }
70
 
71
  def generate():
 
72
  full_response = ""
73
  while True:
74
  conversation = "\n".join([f"{msg['role']}: {msg['content']}" for msg in messages])
 
102
  messages.append({"role": "user", "content": "Please continue your output and do not repeat the previous content"})
103
  break # Jump out of the current loop and continue with the next request
104
  else:
 
105
  last_content = response_message.get('text', '')
106
  if last_content and last_content != full_response:
107
  yield f"data: {json.dumps(format_openai_response(last_content[len(full_response):]))}\n\n"
108
 
109
  yield f"data: {json.dumps(format_openai_response('', finish_reason))}\n\n"
110
  yield "data: [DONE]\n\n"
111
+ return
112
 
 
113
  yield f"data: {json.dumps(format_openai_response('', 'stop'))}\n\n"
114
  yield "data: [DONE]\n\n"
115
 
 
149
  }
150
 
151
  if __name__ == '__main__':
152
+ app.run(debug=True, port=5000)