Update api_usage.py
Browse files- api_usage.py +11 -10
api_usage.py
CHANGED
|
@@ -11,7 +11,7 @@ import asyncio
|
|
| 11 |
import aiohttp
|
| 12 |
|
| 13 |
BASE_URL = 'https://api.openai.com/v1'
|
| 14 |
-
GPT_TYPES = ["gpt-3.5-turbo", "gpt-4", "gpt-4-32k", "gpt-4-32k-0314"]
|
| 15 |
|
| 16 |
TOKEN_LIMIT_PER_TIER_TURBO = {
|
| 17 |
"free": 40000,
|
|
@@ -73,28 +73,28 @@ def get_subscription(key, session, org_list):
|
|
| 73 |
org.append(f"{org_in['id']} ({org_in['name']}, {org_in['title']}, {org_in['role']})")
|
| 74 |
if has_gpt4_32k:
|
| 75 |
list_models_avai.update(GPT_TYPES)
|
| 76 |
-
status_formated = format_status([GPT_TYPES[2], GPT_TYPES[1], GPT_TYPES[0]], session, headers)
|
| 77 |
rpm.append(status_formated[0])
|
| 78 |
tpm.append(status_formated[1])
|
| 79 |
quota.append(status_formated[2])
|
| 80 |
-
list_models.append(f"gpt-4-32k, gpt-4, gpt-3.5-turbo ({len(available_models)} total)")
|
| 81 |
else:
|
| 82 |
list_models_avai.update([GPT_TYPES[3], GPT_TYPES[1], GPT_TYPES[0]])
|
| 83 |
-
status_formated = format_status([GPT_TYPES[3], GPT_TYPES[1], GPT_TYPES[0]], session, headers)
|
| 84 |
rpm.append(status_formated[0])
|
| 85 |
tpm.append(status_formated[1])
|
| 86 |
quota.append(status_formated[2])
|
| 87 |
-
list_models.append(f"gpt-4-32k-0314, gpt-4, gpt-3.5-turbo ({len(available_models)} total)")
|
| 88 |
|
| 89 |
elif has_gpt4:
|
| 90 |
if org_in['id']:
|
| 91 |
org.append(f"{org_in['id']} ({org_in['name']}, {org_in['title']}, {org_in['role']})")
|
| 92 |
list_models_avai.update([GPT_TYPES[1], GPT_TYPES[0]])
|
| 93 |
-
status_formated = format_status([GPT_TYPES[1], GPT_TYPES[0]], session, headers)
|
| 94 |
rpm.append(status_formated[0])
|
| 95 |
tpm.append(status_formated[1])
|
| 96 |
quota.append(status_formated[2])
|
| 97 |
-
list_models.append(f"gpt-4, gpt-3.5-turbo ({len(available_models)} total)")
|
| 98 |
|
| 99 |
else:
|
| 100 |
if org_in['id']:
|
|
@@ -137,7 +137,8 @@ def send_oai_completions(oai_stuff):
|
|
| 137 |
_tpm = '{:,}'.format(tpm_num).replace(',', ' ')
|
| 138 |
_tpm_left = '{:,}'.format(tpm_left).replace(',', ' ')
|
| 139 |
rpm_string = f"{_rpm} ({model})"
|
| 140 |
-
tpm_string = f"{_tpm} ({_tpm_left} left, {model})"
|
|
|
|
| 141 |
dictCount = 0
|
| 142 |
dictLength = len(TOKEN_LIMIT_PER_TIER_GPT4)
|
| 143 |
|
|
@@ -178,8 +179,8 @@ def format_status(list_models_avai, session, headers):
|
|
| 178 |
rpm_str = ""
|
| 179 |
tpm_str = ""
|
| 180 |
for i in range(len(rpm)):
|
| 181 |
-
rpm_str += rpm[i] + ("
|
| 182 |
-
tpm_str += tpm[i] + ("
|
| 183 |
return rpm_str, tpm_str, quota
|
| 184 |
|
| 185 |
def check_key_tier(rpm, tpm, dict, headers):
|
|
|
|
| 11 |
import aiohttp
|
| 12 |
|
| 13 |
BASE_URL = 'https://api.openai.com/v1'
|
| 14 |
+
GPT_TYPES = ["gpt-3.5-turbo", "gpt-4", "gpt-4-32k", "gpt-4-32k-0314", "gpt-4o", "gpt-4-turbo"]
|
| 15 |
|
| 16 |
TOKEN_LIMIT_PER_TIER_TURBO = {
|
| 17 |
"free": 40000,
|
|
|
|
| 73 |
org.append(f"{org_in['id']} ({org_in['name']}, {org_in['title']}, {org_in['role']})")
|
| 74 |
if has_gpt4_32k:
|
| 75 |
list_models_avai.update(GPT_TYPES)
|
| 76 |
+
status_formated = format_status([GPT_TYPES[2], GPT_TYPES[4], GPT_TYPES[5], GPT_TYPES[1], GPT_TYPES[0]], session, headers)
|
| 77 |
rpm.append(status_formated[0])
|
| 78 |
tpm.append(status_formated[1])
|
| 79 |
quota.append(status_formated[2])
|
| 80 |
+
list_models.append(f"gpt-4-32k, gpt-4o, gpt-4-turbo, gpt-4, gpt-3.5-turbo ({len(available_models)} total)")
|
| 81 |
else:
|
| 82 |
list_models_avai.update([GPT_TYPES[3], GPT_TYPES[1], GPT_TYPES[0]])
|
| 83 |
+
status_formated = format_status([GPT_TYPES[3], GPT_TYPES[4], GPT_TYPES[5], GPT_TYPES[1], GPT_TYPES[0]], session, headers)
|
| 84 |
rpm.append(status_formated[0])
|
| 85 |
tpm.append(status_formated[1])
|
| 86 |
quota.append(status_formated[2])
|
| 87 |
+
list_models.append(f"gpt-4-32k-0314, gpt-4o, gpt-4-turbo, gpt-4, gpt-3.5-turbo ({len(available_models)} total)")
|
| 88 |
|
| 89 |
elif has_gpt4:
|
| 90 |
if org_in['id']:
|
| 91 |
org.append(f"{org_in['id']} ({org_in['name']}, {org_in['title']}, {org_in['role']})")
|
| 92 |
list_models_avai.update([GPT_TYPES[1], GPT_TYPES[0]])
|
| 93 |
+
status_formated = format_status([GPT_TYPES[4], GPT_TYPES[5], GPT_TYPES[1], GPT_TYPES[0]], session, headers)
|
| 94 |
rpm.append(status_formated[0])
|
| 95 |
tpm.append(status_formated[1])
|
| 96 |
quota.append(status_formated[2])
|
| 97 |
+
list_models.append(f"gpt-4o, gpt-4-turbo, gpt-4, gpt-3.5-turbo ({len(available_models)} total)")
|
| 98 |
|
| 99 |
else:
|
| 100 |
if org_in['id']:
|
|
|
|
| 137 |
_tpm = '{:,}'.format(tpm_num).replace(',', ' ')
|
| 138 |
_tpm_left = '{:,}'.format(tpm_left).replace(',', ' ')
|
| 139 |
rpm_string = f"{_rpm} ({model})"
|
| 140 |
+
#tpm_string = f"{_tpm} ({_tpm_left} left, {model})"
|
| 141 |
+
tpm_string = f"{_tpm} ({model})"
|
| 142 |
dictCount = 0
|
| 143 |
dictLength = len(TOKEN_LIMIT_PER_TIER_GPT4)
|
| 144 |
|
|
|
|
| 179 |
rpm_str = ""
|
| 180 |
tpm_str = ""
|
| 181 |
for i in range(len(rpm)):
|
| 182 |
+
rpm_str += rpm[i] + (" | " if i < len(rpm)-1 else "")
|
| 183 |
+
tpm_str += tpm[i] + (" | " if i < len(rpm)-1 else "")
|
| 184 |
return rpm_str, tpm_str, quota
|
| 185 |
|
| 186 |
def check_key_tier(rpm, tpm, dict, headers):
|