Upload chatgpt.py
Browse files- Akeno/plugins/chatgpt.py +36 -21
Akeno/plugins/chatgpt.py
CHANGED
@@ -21,8 +21,8 @@ import requests
|
|
21 |
from pyrogram import *
|
22 |
from pyrogram import Client, filters
|
23 |
from pyrogram.types import *
|
24 |
-
from RyuzakiLib import FullStackDev, GeminiLatest, RendyDevChat
|
25 |
-
|
26 |
from Akeno.utils.chat import chat_message
|
27 |
from Akeno.utils.database import db
|
28 |
from Akeno.utils.handler import *
|
@@ -83,12 +83,15 @@ async def googlegm(client: Client, message: Message):
|
|
83 |
& ~filters.forwarded
|
84 |
)
|
85 |
async def chatgpt_images(client: Client, message: Message):
|
86 |
-
|
87 |
-
|
|
|
|
|
|
|
88 |
return await message.reply_text("Give ask from CHATGPT images")
|
89 |
try:
|
90 |
replys = await message.reply_text("Prossing.....")
|
91 |
-
response = await RendyDevChat.image_generator(
|
92 |
x = response["randydev"].get("url")
|
93 |
for i, url in enumerate(x, start=1):
|
94 |
await FullStackDev.fast(url, filename=f"original_{i}.png")
|
@@ -112,15 +115,18 @@ async def chatgpt_images(client: Client, message: Message):
|
|
112 |
& ~filters.forwarded
|
113 |
)
|
114 |
async def faceai_(client: Client, message: Message):
|
115 |
-
|
116 |
-
|
117 |
-
|
|
|
|
|
|
|
118 |
try:
|
119 |
clients_name, token = await db.get_env(ENV_TEMPLATE.face_clients_name), await db.get_env(ENV_TEMPLATE.face_token_key)
|
120 |
if not clients_name and not token:
|
121 |
return await message.reply_text("Required .setvar FACE_CLIENTS_NAME xxxx and .setvar FACE_TOKEN xxxx")
|
122 |
send = FaceAI(clients_name=clients_name, token=token)
|
123 |
-
response = await send.chat(
|
124 |
if len(response) > 4096:
|
125 |
with open("chat.txt", "w+", encoding="utf8") as out_file:
|
126 |
out_file.write(response)
|
@@ -142,11 +148,14 @@ async def faceai_(client: Client, message: Message):
|
|
142 |
& ~filters.forwarded
|
143 |
)
|
144 |
async def mistralai_(client: Client, message: Message):
|
145 |
-
|
146 |
-
|
147 |
-
|
|
|
|
|
|
|
148 |
try:
|
149 |
-
messager = await mistraai(
|
150 |
if messager is None:
|
151 |
return await message.reply_text("No response")
|
152 |
output = messager["randydev"].get("message")
|
@@ -171,11 +180,14 @@ async def mistralai_(client: Client, message: Message):
|
|
171 |
& ~filters.forwarded
|
172 |
)
|
173 |
async def chatgpt_old_(client: Client, message: Message):
|
174 |
-
|
175 |
-
|
176 |
-
|
|
|
|
|
|
|
177 |
try:
|
178 |
-
messager = await chatgptold(
|
179 |
if messager is None:
|
180 |
return await message.reply_text("No response")
|
181 |
output = messager["randydev"].get("message")
|
@@ -200,11 +212,14 @@ async def chatgpt_old_(client: Client, message: Message):
|
|
200 |
& ~filters.forwarded
|
201 |
)
|
202 |
async def chatgpt(client: Client, message: Message):
|
203 |
-
|
204 |
-
|
205 |
-
|
|
|
|
|
|
|
206 |
try:
|
207 |
-
messager = await chat_message(
|
208 |
if len(messager) > 4096:
|
209 |
with open("chat.txt", "w+", encoding="utf8") as out_file:
|
210 |
out_file.write(messager)
|
|
|
21 |
from pyrogram import *
|
22 |
from pyrogram import Client, filters
|
23 |
from pyrogram.types import *
|
24 |
+
from RyuzakiLib import FaceAI, FullStackDev, GeminiLatest, RendyDevChat
|
25 |
+
|
26 |
from Akeno.utils.chat import chat_message
|
27 |
from Akeno.utils.database import db
|
28 |
from Akeno.utils.handler import *
|
|
|
83 |
& ~filters.forwarded
|
84 |
)
|
85 |
async def chatgpt_images(client: Client, message: Message):
|
86 |
+
if len(message.command) > 1:
|
87 |
+
prompt = message.text.split(maxsplit=1)[1]
|
88 |
+
elif message.reply_to_message:
|
89 |
+
prompt = message.reply_to_message.text
|
90 |
+
else:
|
91 |
return await message.reply_text("Give ask from CHATGPT images")
|
92 |
try:
|
93 |
replys = await message.reply_text("Prossing.....")
|
94 |
+
response = await RendyDevChat.image_generator(prompt)
|
95 |
x = response["randydev"].get("url")
|
96 |
for i, url in enumerate(x, start=1):
|
97 |
await FullStackDev.fast(url, filename=f"original_{i}.png")
|
|
|
115 |
& ~filters.forwarded
|
116 |
)
|
117 |
async def faceai_(client: Client, message: Message):
|
118 |
+
if len(message.command) > 1:
|
119 |
+
prompt = message.text.split(maxsplit=1)[1]
|
120 |
+
elif message.reply_to_message:
|
121 |
+
prompt = message.reply_to_message.text
|
122 |
+
else:
|
123 |
+
return await message.reply_text("Give ask from FaceAI")
|
124 |
try:
|
125 |
clients_name, token = await db.get_env(ENV_TEMPLATE.face_clients_name), await db.get_env(ENV_TEMPLATE.face_token_key)
|
126 |
if not clients_name and not token:
|
127 |
return await message.reply_text("Required .setvar FACE_CLIENTS_NAME xxxx and .setvar FACE_TOKEN xxxx")
|
128 |
send = FaceAI(clients_name=clients_name, token=token)
|
129 |
+
response = await send.chat(prompt, no_db=True)
|
130 |
if len(response) > 4096:
|
131 |
with open("chat.txt", "w+", encoding="utf8") as out_file:
|
132 |
out_file.write(response)
|
|
|
148 |
& ~filters.forwarded
|
149 |
)
|
150 |
async def mistralai_(client: Client, message: Message):
|
151 |
+
if len(message.command) > 1:
|
152 |
+
prompt = message.text.split(maxsplit=1)[1]
|
153 |
+
elif message.reply_to_message:
|
154 |
+
prompt = message.reply_to_message.text
|
155 |
+
else:
|
156 |
+
return await message.reply_text("Give ask from mistralai")
|
157 |
try:
|
158 |
+
messager = await mistraai(prompt)
|
159 |
if messager is None:
|
160 |
return await message.reply_text("No response")
|
161 |
output = messager["randydev"].get("message")
|
|
|
180 |
& ~filters.forwarded
|
181 |
)
|
182 |
async def chatgpt_old_(client: Client, message: Message):
|
183 |
+
if len(message.command) > 1:
|
184 |
+
prompt = message.text.split(maxsplit=1)[1]
|
185 |
+
elif message.reply_to_message:
|
186 |
+
prompt = message.reply_to_message.text
|
187 |
+
else:
|
188 |
+
return await message.reply_text("Give ask from CHATGPT-3")
|
189 |
try:
|
190 |
+
messager = await chatgptold(prompt)
|
191 |
if messager is None:
|
192 |
return await message.reply_text("No response")
|
193 |
output = messager["randydev"].get("message")
|
|
|
212 |
& ~filters.forwarded
|
213 |
)
|
214 |
async def chatgpt(client: Client, message: Message):
|
215 |
+
if len(message.command) > 1:
|
216 |
+
prompt = message.text.split(maxsplit=1)[1]
|
217 |
+
elif message.reply_to_message:
|
218 |
+
prompt = message.reply_to_message.text
|
219 |
+
else:
|
220 |
+
return await message.reply_text("Give ask from CHATGPT-4O")
|
221 |
try:
|
222 |
+
messager = await chat_message(prompt)
|
223 |
if len(messager) > 4096:
|
224 |
with open("chat.txt", "w+", encoding="utf8") as out_file:
|
225 |
out_file.write(messager)
|