From f1594cfddecf2bd47a25d54852834b392eaa4efd Mon Sep 17 00:00:00 2001 From: Aymane Hrouch Date: Thu, 27 Apr 2023 15:43:59 +0100 Subject: Reformat code using PyCharm --- unfinished/bard/__init__.py | 31 +++--- unfinished/bard/typings.py | 2 +- unfinished/bing/__ini__.py | 39 ++++---- unfinished/cocalc/__init__.py | 1 - unfinished/cocalc/cocalc_test.py | 5 +- unfinished/easyai/main.py | 5 +- unfinished/gptbz/__init__.py | 4 +- unfinished/openai/__ini__.py | 3 +- unfinished/openaihosted/__init__.py | 5 +- unfinished/openprompt/create.py | 20 ++-- unfinished/openprompt/mail.py | 6 +- unfinished/openprompt/main.py | 5 +- unfinished/openprompt/test.py | 5 +- unfinished/t3nsor/__init__.py | 131 ++++++++++++------------ unfinished/test.py | 8 +- unfinished/theb.ai/__init__.py | 45 +++++---- unfinished/vercelai/v2.py | 6 +- unfinished/writesonic/__init__.py | 195 ++++++++++++++++++------------------ 18 files changed, 262 insertions(+), 254 deletions(-) (limited to 'unfinished') diff --git a/unfinished/bard/__init__.py b/unfinished/bard/__init__.py index ef8980d7..f1d68b92 100644 --- a/unfinished/bard/__init__.py +++ b/unfinished/bard/__init__.py @@ -1,12 +1,12 @@ -from requests import Session -from re import search -from random import randint from json import dumps, loads -from urllib.parse import urlencode -from dotenv import load_dotenv from os import getenv +from random import randint +from re import search +from urllib.parse import urlencode from bard.typings import BardResponse +from dotenv import load_dotenv +from requests import Session load_dotenv() token = getenv('1psid') @@ -62,16 +62,17 @@ class Completion: 'rt': 'c', }) - response = client.post(f'https://bard.google.com/_/BardChatUi/data/assistant.lamda.BardFrontendService/StreamGenerate?{params}', - data={ - 'at': snlm0e, - 'f.req': dumps([None, dumps([ - [prompt], - None, - [conversation_id, response_id, choice_id], - ])]) - } - ) + response = client.post( + f'https://bard.google.com/_/BardChatUi/data/assistant.lamda.BardFrontendService/StreamGenerate?{params}', + data={ + 'at': snlm0e, + 'f.req': dumps([None, dumps([ + [prompt], + None, + [conversation_id, response_id, choice_id], + ])]) + } + ) chat_data = loads(response.content.splitlines()[3])[0][2] if not chat_data: diff --git a/unfinished/bard/typings.py b/unfinished/bard/typings.py index ddf803b6..75b73bf9 100644 --- a/unfinished/bard/typings.py +++ b/unfinished/bard/typings.py @@ -1,4 +1,4 @@ -from typing import Dict, List, Optional, Union +from typing import Dict, List, Union class BardResponse: diff --git a/unfinished/bing/__ini__.py b/unfinished/bing/__ini__.py index 508b1067..1e4fd149 100644 --- a/unfinished/bing/__ini__.py +++ b/unfinished/bing/__ini__.py @@ -1,14 +1,12 @@ # Import necessary libraries -from requests import get -from browser_cookie3 import edge, chrome -from ssl import create_default_context -from certifi import where -from uuid import uuid4 -from random import randint +import asyncio from json import dumps, loads +from ssl import create_default_context -import asyncio import websockets +from browser_cookie3 import edge +from certifi import where +from requests import get # Set up SSL context ssl_context = create_default_context() @@ -28,14 +26,14 @@ def get_token(): class AsyncCompletion: async def create( - prompt: str = 'hello world', - optionSets: list = [ - 'deepleo', - 'enable_debug_commands', - 'disable_emoji_spoken_text', - 'enablemm', - 'h3relaxedimg' - ], + prompt: str = 'hello world', + optionSets: list = [ + 'deepleo', + 'enable_debug_commands', + 'disable_emoji_spoken_text', + 'enablemm', + 'h3relaxedimg' + ], token: str = get_token()): """Create a connection to Bing AI and send the prompt.""" @@ -83,7 +81,7 @@ class AsyncCompletion: continue response = loads(obj) - if response.get('type') == 1 and response['arguments'][0].get('messages',): + if response.get('type') == 1 and response['arguments'][0].get('messages', ): response_text = response['arguments'][0]['messages'][0]['adaptiveCards'][0]['body'][0].get( 'text') @@ -99,11 +97,12 @@ class AsyncCompletion: async def run(): """Run the async completion and print the result.""" async for value in AsyncCompletion.create( - prompt='summarize cinderella with each word beginning with a consecutive letter of the alphabet, a-z', - optionSets=[ - "galileo", - ] + prompt='summarize cinderella with each word beginning with a consecutive letter of the alphabet, a-z', + optionSets=[ + "galileo", + ] ): print(value, end='', flush=True) + asyncio.run(run()) diff --git a/unfinished/cocalc/__init__.py b/unfinished/cocalc/__init__.py index 3ad9937a..2b73fc9e 100644 --- a/unfinished/cocalc/__init__.py +++ b/unfinished/cocalc/__init__.py @@ -6,7 +6,6 @@ class Completion: system_prompt=("ASSUME I HAVE FULL ACCESS TO COCALC. ENCLOSE MATH IN $. " "INCLUDE THE LANGUAGE DIRECTLY AFTER THE TRIPLE BACKTICKS " "IN ALL MARKDOWN CODE BLOCKS. How can I do the following using CoCalc?")) -> str: - # Initialize a session with custom headers session = self._initialize_session() diff --git a/unfinished/cocalc/cocalc_test.py b/unfinished/cocalc/cocalc_test.py index 0e1a7b3b..d6266518 100644 --- a/unfinished/cocalc/cocalc_test.py +++ b/unfinished/cocalc/cocalc_test.py @@ -1,8 +1,7 @@ import cocalc - response = cocalc.Completion.create( - prompt = 'hello world' + prompt='hello world' ) -print(response) \ No newline at end of file +print(response) diff --git a/unfinished/easyai/main.py b/unfinished/easyai/main.py index 07adfd72..451adb3b 100644 --- a/unfinished/easyai/main.py +++ b/unfinished/easyai/main.py @@ -1,7 +1,8 @@ # Import necessary libraries -from requests import get -from os import urandom from json import loads +from os import urandom + +from requests import get # Generate a random session ID sessionId = urandom(10).hex() diff --git a/unfinished/gptbz/__init__.py b/unfinished/gptbz/__init__.py index 44bfcd19..e95d5716 100644 --- a/unfinished/gptbz/__init__.py +++ b/unfinished/gptbz/__init__.py @@ -1,6 +1,8 @@ -import websockets from json import dumps, loads +import websockets + + # Define the asynchronous function to test the WebSocket connection diff --git a/unfinished/openai/__ini__.py b/unfinished/openai/__ini__.py index 71ec4623..f0894e1b 100644 --- a/unfinished/openai/__ini__.py +++ b/unfinished/openai/__ini__.py @@ -1,7 +1,8 @@ # Import required libraries -from tls_client import Session from uuid import uuid4 + from browser_cookie3 import chrome +from tls_client import Session class OpenAIChat: diff --git a/unfinished/openaihosted/__init__.py b/unfinished/openaihosted/__init__.py index e857b475..ba4d3982 100644 --- a/unfinished/openaihosted/__init__.py +++ b/unfinished/openaihosted/__init__.py @@ -1,7 +1,8 @@ -import requests import json import re +import requests + headers = { 'authority': 'openai.a2hosted.com', 'accept': 'text/event-stream', @@ -13,10 +14,12 @@ headers = { 'user-agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36 Edg/113.0.0.0', } + def create_query_param(conversation): encoded_conversation = json.dumps(conversation) return encoded_conversation.replace(" ", "%20").replace('"', '%22').replace("'", "%27") + user_input = input("Enter your message: ") data = [ diff --git a/unfinished/openprompt/create.py b/unfinished/openprompt/create.py index 6ccb66f4..c968c162 100644 --- a/unfinished/openprompt/create.py +++ b/unfinished/openprompt/create.py @@ -1,9 +1,9 @@ -from requests import post, get from json import dumps -#from mail import MailClient -from time import sleep +# from mail import MailClient from re import findall +from requests import post, get + html = get('https://developermail.com/mail/') print(html.cookies.get('mailboxId')) email = findall(r'mailto:(.*)">', html.text)[0] @@ -15,9 +15,9 @@ headers = { } json_data = { - 'email' : email, + 'email': email, 'password': 'T4xyt4Yn6WWQ4NC', - 'data' : {}, + 'data': {}, 'gotrue_meta_security': {}, } @@ -27,20 +27,20 @@ print(response.json()) # email_link = None # while not email_link: # sleep(1) - + # mails = mailbox.getmails() # print(mails) quit() -url = input("Enter the url: ") +url = input("Enter the url: ") response = get(url, allow_redirects=False) # https://openprompt.co/#access_token=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJhdWQiOiJhdXRoZW50aWNhdGVkIiwiZXhwIjoxNjgyMjk0ODcxLCJzdWIiOiI4NWNkNTNiNC1lZTUwLTRiMDQtOGJhNS0wNTUyNjk4ODliZDIiLCJlbWFpbCI6ImNsc2J5emdqcGhiQGJ1Z2Zvby5jb20iLCJwaG9uZSI6IiIsImFwcF9tZXRhZGF0YSI6eyJwcm92aWRlciI6ImVtYWlsIiwicHJvdmlkZXJzIjpbImVtYWlsIl19LCJ1c2VyX21ldGFkYXRhIjp7fSwicm9sZSI6ImF1dGhlbnRpY2F0ZWQiLCJhYWwiOiJhYWwxIiwiYW1yIjpbeyJtZXRob2QiOiJvdHAiLCJ0aW1lc3RhbXAiOjE2ODE2OTAwNzF9XSwic2Vzc2lvbl9pZCI6ImY4MTg1YTM5LTkxYzgtNGFmMy1iNzAxLTdhY2MwY2MwMGNlNSJ9.UvcTfpyIM1TdzM8ZV6UAPWfa0rgNq4AiqeD0INy6zV8&expires_in=604800&refresh_token=_Zp8uXIA2InTDKYgo8TCqA&token_type=bearer&type=signup -redirect = response.headers.get('location') -access_token = redirect.split('&')[0].split('=')[1] +redirect = response.headers.get('location') +access_token = redirect.split('&')[0].split('=')[1] refresh_token = redirect.split('&')[2].split('=')[1] supabase_auth_token = dumps([access_token, refresh_token, None, None, None], separators=(',', ':')) @@ -61,4 +61,4 @@ json_data = { response = post('https://openprompt.co/api/chat2', cookies=cookies, json=json_data, stream=True) for chunk in response.iter_content(chunk_size=1024): - print(chunk) \ No newline at end of file + print(chunk) diff --git a/unfinished/openprompt/mail.py b/unfinished/openprompt/mail.py index 082ac9fb..1130e7df 100644 --- a/unfinished/openprompt/mail.py +++ b/unfinished/openprompt/mail.py @@ -1,6 +1,8 @@ -import requests import email +import requests + + class MailClient: def __init__(self): @@ -106,4 +108,4 @@ class MailClient: client = MailClient() client.newtoken() -print(client.getmails()) \ No newline at end of file +print(client.getmails()) diff --git a/unfinished/openprompt/main.py b/unfinished/openprompt/main.py index 2fa4508e..e68a3b63 100644 --- a/unfinished/openprompt/main.py +++ b/unfinished/openprompt/main.py @@ -30,8 +30,7 @@ json_data = { ], } -response = requests.post('https://openprompt.co/api/chat2', cookies=cookies, headers=headers, json=json_data, stream=True) +response = requests.post('https://openprompt.co/api/chat2', cookies=cookies, headers=headers, json=json_data, + stream=True) for chunk in response.iter_content(chunk_size=1024): print(chunk) - - diff --git a/unfinished/openprompt/test.py b/unfinished/openprompt/test.py index d178462e..65319cb6 100644 --- a/unfinished/openprompt/test.py +++ b/unfinished/openprompt/test.py @@ -1,7 +1,6 @@ access_token = 'eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJhdWQiOiJhdXRoZW50aWNhdGVkIiwiZXhwIjoxNjgyMjk0ODcxLCJzdWIiOiI4NWNkNTNiNC1lZTUwLTRiMDQtOGJhNS0wNTUyNjk4ODliZDIiLCJlbWFpbCI6ImNsc2J5emdqcGhiQGJ1Z2Zvby5jb20iLCJwaG9uZSI6IiIsImFwcF9tZXRhZGF0YSI6eyJwcm92aWRlciI6ImVtYWlsIiwicHJvdmlkZXJzIjpbImVtYWlsIl19LCJ1c2VyX21ldGFkYXRhIjp7fSwicm9sZSI6ImF1dGhlbnRpY2F0ZWQiLCJhYWwiOiJhYWwxIiwiYW1yIjpbeyJtZXRob2QiOiJvdHAiLCJ0aW1lc3RhbXAiOjE2ODE2OTAwNzF9XSwic2Vzc2lvbl9pZCI6ImY4MTg1YTM5LTkxYzgtNGFmMy1iNzAxLTdhY2MwY2MwMGNlNSJ9.UvcTfpyIM1TdzM8ZV6UAPWfa0rgNq4AiqeD0INy6zV' -supabase_auth_token= '%5B%22eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJhdWQiOiJhdXRoZW50aWNhdGVkIiwiZXhwIjoxNjgyMjk0ODcxLCJzdWIiOiI4NWNkNTNiNC1lZTUwLTRiMDQtOGJhNS0wNTUyNjk4ODliZDIiLCJlbWFpbCI6ImNsc2J5emdqcGhiQGJ1Z2Zvby5jb20iLCJwaG9uZSI6IiIsImFwcF9tZXRhZGF0YSI6eyJwcm92aWRlciI6ImVtYWlsIiwicHJvdmlkZXJzIjpbImVtYWlsIl19LCJ1c2VyX21ldGFkYXRhIjp7fSwicm9sZSI6ImF1dGhlbnRpY2F0ZWQiLCJhYWwiOiJhYWwxIiwiYW1yIjpbeyJtZXRob2QiOiJvdHAiLCJ0aW1lc3RhbXAiOjE2ODE2OTAwNzF9XSwic2Vzc2lvbl9pZCI6ImY4MTg1YTM5LTkxYzgtNGFmMy1iNzAxLTdhY2MwY2MwMGNlNSJ9.UvcTfpyIM1TdzM8ZV6UAPWfa0rgNq4AiqeD0INy6zV8%22%2C%22_Zp8uXIA2InTDKYgo8TCqA%22%2Cnull%2Cnull%2Cnull%5D' - +supabase_auth_token = '%5B%22eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJhdWQiOiJhdXRoZW50aWNhdGVkIiwiZXhwIjoxNjgyMjk0ODcxLCJzdWIiOiI4NWNkNTNiNC1lZTUwLTRiMDQtOGJhNS0wNTUyNjk4ODliZDIiLCJlbWFpbCI6ImNsc2J5emdqcGhiQGJ1Z2Zvby5jb20iLCJwaG9uZSI6IiIsImFwcF9tZXRhZGF0YSI6eyJwcm92aWRlciI6ImVtYWlsIiwicHJvdmlkZXJzIjpbImVtYWlsIl19LCJ1c2VyX21ldGFkYXRhIjp7fSwicm9sZSI6ImF1dGhlbnRpY2F0ZWQiLCJhYWwiOiJhYWwxIiwiYW1yIjpbeyJtZXRob2QiOiJvdHAiLCJ0aW1lc3RhbXAiOjE2ODE2OTAwNzF9XSwic2Vzc2lvbl9pZCI6ImY4MTg1YTM5LTkxYzgtNGFmMy1iNzAxLTdhY2MwY2MwMGNlNSJ9.UvcTfpyIM1TdzM8ZV6UAPWfa0rgNq4AiqeD0INy6zV8%22%2C%22_Zp8uXIA2InTDKYgo8TCqA%22%2Cnull%2Cnull%2Cnull%5D' idk = [ "eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJhdWQiOiJhdXRoZW50aWNhdGVkIiwiZXhwIjoxNjgyMjk0ODcxLCJzdWIiOiI4NWNkNTNiNC1lZTUwLTRiMDQtOGJhNS0wNTUyNjk4ODliZDIiLCJlbWFpbCI6ImNsc2J5emdqcGhiQGJ1Z2Zvby5jb20iLCJwaG9uZSI6IiIsImFwcF9tZXRhZGF0YSI6eyJwcm92aWRlciI6ImVtYWlsIiwicHJvdmlkZXJzIjpbImVtYWlsIl19LCJ1c2VyX21ldGFkYXRhIjp7fSwicm9sZSI6ImF1dGhlbnRpY2F0ZWQiLCJhYWwiOiJhYWwxIiwiYW1yIjpbeyJtZXRob2QiOiJvdHAiLCJ0aW1lc3RhbXAiOjE2ODE2OTAwNzF9XSwic2Vzc2lvbl9pZCI6ImY4MTg1YTM5LTkxYzgtNGFmMy1iNzAxLTdhY2MwY2MwMGNlNSJ9.UvcTfpyIM1TdzM8ZV6UAPWfa0rgNq4AiqeD0INy6zV8", - "_Zp8uXIA2InTDKYgo8TCqA",None,None,None] \ No newline at end of file + "_Zp8uXIA2InTDKYgo8TCqA", None, None, None] diff --git a/unfinished/t3nsor/__init__.py b/unfinished/t3nsor/__init__.py index aec45dcf..9b588e98 100644 --- a/unfinished/t3nsor/__init__.py +++ b/unfinished/t3nsor/__init__.py @@ -1,5 +1,6 @@ +from time import time + from requests import post -from time import time headers = { 'authority': 'www.t3nsor.tech', @@ -19,18 +20,17 @@ headers = { 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36', } + class T3nsorResponse: - class Completion: - class Choices: def __init__(self, choice: dict) -> None: - self.text = choice['text'] - self.content = self.text.encode() - self.index = choice['index'] - self.logprobs = choice['logprobs'] - self.finish_reason = choice['finish_reason'] - + self.text = choice['text'] + self.content = self.text.encode() + self.index = choice['index'] + self.logprobs = choice['logprobs'] + self.finish_reason = choice['finish_reason'] + def __repr__(self) -> str: return f'''<__main__.APIResponse.Completion.Choices(\n text = {self.text.encode()},\n index = {self.index},\n logprobs = {self.logprobs},\n finish_reason = {self.finish_reason})object at 0x1337>''' @@ -39,99 +39,98 @@ class T3nsorResponse: class Usage: def __init__(self, usage_dict: dict) -> None: - self.prompt_tokens = usage_dict['prompt_chars'] - self.completion_tokens = usage_dict['completion_chars'] - self.total_tokens = usage_dict['total_chars'] + self.prompt_tokens = usage_dict['prompt_chars'] + self.completion_tokens = usage_dict['completion_chars'] + self.total_tokens = usage_dict['total_chars'] def __repr__(self): return f'''<__main__.APIResponse.Usage(\n prompt_tokens = {self.prompt_tokens},\n completion_tokens = {self.completion_tokens},\n total_tokens = {self.total_tokens})object at 0x1337>''' - + def __init__(self, response_dict: dict) -> None: - - self.response_dict = response_dict - self.id = response_dict['id'] - self.object = response_dict['object'] - self.created = response_dict['created'] - self.model = response_dict['model'] - self.completion = self.Completion(response_dict['choices']) - self.usage = self.Usage(response_dict['usage']) + self.response_dict = response_dict + self.id = response_dict['id'] + self.object = response_dict['object'] + self.created = response_dict['created'] + self.model = response_dict['model'] + self.completion = self.Completion(response_dict['choices']) + self.usage = self.Usage(response_dict['usage']) def json(self) -> dict: return self.response_dict + class Completion: model = { 'model': { - 'id' : 'gpt-3.5-turbo', - 'name' : 'Default (GPT-3.5)' + 'id': 'gpt-3.5-turbo', + 'name': 'Default (GPT-3.5)' } } def create( - prompt: str = 'hello world', - messages: list = []) -> T3nsorResponse: - - response = post('https://www.t3nsor.tech/api/chat', headers = headers, json = Completion.model | { - 'messages' : messages, - 'key' : '', - 'prompt' : prompt + prompt: str = 'hello world', + messages: list = []) -> T3nsorResponse: + response = post('https://www.t3nsor.tech/api/chat', headers=headers, json=Completion.model | { + 'messages': messages, + 'key': '', + 'prompt': prompt }) return T3nsorResponse({ - 'id' : f'cmpl-1337-{int(time())}', - 'object' : 'text_completion', - 'created': int(time()), - 'model' : Completion.model, + 'id': f'cmpl-1337-{int(time())}', + 'object': 'text_completion', + 'created': int(time()), + 'model': Completion.model, 'choices': [{ - 'text' : response.text, - 'index' : 0, - 'logprobs' : None, - 'finish_reason' : 'stop' - }], + 'text': response.text, + 'index': 0, + 'logprobs': None, + 'finish_reason': 'stop' + }], 'usage': { - 'prompt_chars' : len(prompt), - 'completion_chars' : len(response.text), - 'total_chars' : len(prompt) + len(response.text) + 'prompt_chars': len(prompt), + 'completion_chars': len(response.text), + 'total_chars': len(prompt) + len(response.text) } }) + class StreamCompletion: model = { 'model': { - 'id' : 'gpt-3.5-turbo', - 'name' : 'Default (GPT-3.5)' + 'id': 'gpt-3.5-turbo', + 'name': 'Default (GPT-3.5)' } } def create( - prompt: str = 'hello world', - messages: list = []) -> T3nsorResponse: - + prompt: str = 'hello world', + messages: list = []) -> T3nsorResponse: print('t3nsor api is down, this may not work, refer to another module') - response = post('https://www.t3nsor.tech/api/chat', headers = headers, stream = True, json = Completion.model | { - 'messages' : messages, - 'key' : '', - 'prompt' : prompt + response = post('https://www.t3nsor.tech/api/chat', headers=headers, stream=True, json=Completion.model | { + 'messages': messages, + 'key': '', + 'prompt': prompt }) - - for chunk in response.iter_content(chunk_size = 2046): + + for chunk in response.iter_content(chunk_size=2046): yield T3nsorResponse({ - 'id' : f'cmpl-1337-{int(time())}', - 'object' : 'text_completion', - 'created': int(time()), - 'model' : Completion.model, - + 'id': f'cmpl-1337-{int(time())}', + 'object': 'text_completion', + 'created': int(time()), + 'model': Completion.model, + 'choices': [{ - 'text' : chunk.decode(), - 'index' : 0, - 'logprobs' : None, - 'finish_reason' : 'stop' + 'text': chunk.decode(), + 'index': 0, + 'logprobs': None, + 'finish_reason': 'stop' }], - + 'usage': { - 'prompt_chars' : len(prompt), - 'completion_chars' : len(chunk.decode()), - 'total_chars' : len(prompt) + len(chunk.decode()) + 'prompt_chars': len(prompt), + 'completion_chars': len(chunk.decode()), + 'total_chars': len(prompt) + len(chunk.decode()) } }) diff --git a/unfinished/test.py b/unfinished/test.py index 93e39cc9..a5f038c5 100644 --- a/unfinished/test.py +++ b/unfinished/test.py @@ -1,12 +1,8 @@ -import gptbz -import asyncio - - # asyncio.run(gptbz.test()) import requests image = '/9j/4AAQSkZJRgABAQEAYABgAAD/2wBDAAgGBgcGBQgHBwcJCQgKDBQNDAsLDBkSEw8UHRofHh0aHBwgJC4nICIsIxwcKDcpLDAxNDQ0Hyc5PTgyPC4zNDL/2wBDAQkJCQwLDBgNDRgyIRwhMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjL/wAARCAAoALQDASIAAhEBAxEB/8QAHwAAAQUBAQEBAQEAAAAAAAAAAAECAwQFBgcICQoL/8QAtRAAAgEDAwIEAwUFBAQAAAF9AQIDAAQRBRIhMUEGE1FhByJxFDKBkaEII0KxwRVS0fAkM2JyggkKFhcYGRolJicoKSo0NTY3ODk6Q0RFRkdISUpTVFVWV1hZWmNkZWZnaGlqc3R1dnd4eXqDhIWGh4iJipKTlJWWl5iZmqKjpKWmp6ipqrKztLW2t7i5usLDxMXGx8jJytLT1NXW19jZ2uHi4+Tl5ufo6erx8vP09fb3+Pn6/8QAHwEAAwEBAQEBAQEBAQAAAAAAAAECAwQFBgcICQoL/8QAtREAAgECBAQDBAcFBAQAAQJ3AAECAxEEBSExBhJBUQdhcRMiMoEIFEKRobHBCSMzUvAVYnLRChYkNOEl8RcYGRomJygpKjU2Nzg5OkNERUZHSElKU1RVVldYWVpjZGVmZ2hpanN0dXZ3eHl6goOEhYaHiImKkpOUlZaXmJmaoqOkpaanqKmqsrO0tba3uLm6wsPExcbHyMnK0tPU1dbX2Nna4uPk5ebn6Onq8vP09fb3+Pn6/9oADAMBAAIRAxEAPwD3+iiigDkZP+EhS4W0k1S+VntQPtEWmRsgkNwBu4ZsHYQNvTbls5BA6DS7uW6S6E0VwjQ3UsQM0Pl71DZUrydy4IAbvg8CsTx3DbHQLi4uVs9scWzdd+dsAaWI4PlfNjKjpzkDtmpoNSgbWYpLR7Ty5bq5trw/vd3nIowBxtzti53Y6fKT3z2djra56fNbv07HR1z13ZRX/jDyby0+02f9nfdmsEeHd5o/5anndwPkxjjPWuhrh9Mvra88RLqccmnOHtvLEqfaN+1r1lUcjbg4PbO4H+Cqk+hnRi9ZI29E0uC2N1eG3Am+13DITZRwuqlsYG0ZYEKCGJywwT2AtWTapcW1vcPPCiyrE5ils2SRQV+dW/ecMT/3zgj5utZtpdwL4e190e02W9xeb9vm7FOWY78/NnnJ28f3ahkgtptD8JRlbMos9s8QPnbcrEzDy/4sgDjzOMdeaSZbi23f8vmbfn6hBFuktmuWWPJWCNELNuxgbpcDj1Pbr2qJ9bMVyIZNK1JVLyr5qwB1AjUNu+Uk4bovGSRjAqCTwdoElv5B02MReT5G1HZfk8zzMcEfx81YlsJ7NJX0tolZzNK8dyZJA8jDIwd3yjcBkAHjOAM09SP3b/q36mkjiSNXAYBgCNykH8QeRWdfaw1ldSW66XqN0UgE++3iBRsvt2BiQN/8WPQZqharF9oN5osVml1NLbLqUbmUFY/L4CrgYYKy4yoGM5xjhlnc2OoeMrfULV7aQXGkExyYlErJ5oPQ/Jtye/zZ9qLgqaTba0NyzvPtizH7NcQeVM8OJ49u/acbl9VPY96s1geFjF/xOhF9m41Wfd9n8z73BO7f/Fzzt+X0q7c6mWvRY2DwSXcUsQuUff8Auo2ySflB+YqrYyQOmTyARPQmVP32kLqF1cbmsrJZkuni3rcfZ98UfzKvJJUE4JOM5wpODwDl3Meuf2rHbRatcBJXuj5iachjhUovlBmZudrNkEZ3HIOMGlhREhbS9He2a8MO6a4fzmGDMQ3zAk5yZ8DzMgj0yRuWdha2CzLawrEJpnnkx/G7HLMfc0bl3VNf5pff/kVLS8uxFHHJZ3s5Xyo2mZI4y2VBZyN44B6gDrwAcVZ069Go2EV2Le5t/MBPlXMZjkXnGGU9OlULSdbfTt8LWy5mt0JAkK4YRLjnnODx26Z71TXULEWn/CUWDwmxeDbM4WbkCXJbaB23SnlM5PUDNF7CcObZf12OlpCcDoTz2oVlcZVgRkjIPccGo7hgsSk7ceYg+bP94elUYpamda64915GdH1SESxiTM0KjZmTZtbDHB53Y/u89eK1qw4xD9l0mIC3wLdCg/eYwHh+73x0+9znb71uUkXUSWyCiiimZhRRRQBieL5Hj8LXjxySxuNmGivFtWHzr0lbhfx69O9MvHdZpbKKWYnUluNji+VGikVFULHnkdGbjO05JHPEviyF5/DF7HGkjuQpCx2i3THDA8RNw3Tv069qR0kk0i4uFilF3bSXTwE2a+YGzIAUQnnIPByN46kbjUPc6YNKC9X+SLtjeB9Mt5ZyqzbI1lQzK5R2C/KWGAT8w6dcjHUVzemSyxeCba9e5uWfzIgxl1aOTgXPebGw5BwR3ACdalna8+0R3Kx3nk6jc2MvkjTI2MH97zDnI+4uWOSny4z2Lqxmt/hytvHHIZhFHJsj0yJnyXDEfZ87M9cjPB56ik2y4xSsu7XcnjMsejeJszXBZZrgozaihZAYwQFfGIQM8Bvu9ehrTKuJtOg3y5gKs/8ApAy2Y5B846uMj8Tz/CaqzROH1C3EchW6uHGRZIVx9nHXs4yPvN1PydBV2Lc+u3eUkCJBDtZoAFJzJna/VjgjI/h/4EaaM5PS/wDXRF+iiirOcy7RZE8RanukmKPFA6q9yHVfvg7Y+qfd5J4Y9OhrJ8Nm4FxYJNNdORaXCsJtTS4yVnAyQoG5sfxfw/dPJrUslmGt6rcymQxM0MMStahMALk4cfM65c9cBSGA7mqmi2k9t/ZZuDJJKbSdpHNjHEdzyRvhtv3G5PyjIbBJOVqDpurP5d+zGWtzeLdahZQLNK895PiV7+N/IURKQQMEqNzKAm1tucnggG4Fkhs4INNuJL145oEuHa7BcIAuWOQRkrhiAFzkkEE8rNDJPczWtnG1rG7yfapvsqESsY1AIJPP3hztbPllTjHKvpv2CWKbTUSHdJCk8cVtH+8jUFOSNpGAynOTgJgL1BNRNxf9fmWNGa3fR7U2ty9zDswJZJxMzHvlwSCc5BwccVerBZ3tLf8Atqyguvsxt/n02OyUSsxk3FsHa24bnyM4ycgE9d1WDDIz1I5BHQ471SM6i1uY8cjjSIWLyFjLbDJu1J5Mefn6HryP4snH3hRdmTS5f7T82aS2WBY5Y5LpVjX94Pn+YYzhmydw4UDB4wio/wDY8K+XLuE1qcfY1B4MWfk6DHOT/Bg4+6K1zGkkHlSoroy7WVlGCCOQRSsU5JGUrPo96EZ5p7O7mmmlubm7XFqQoYIobB2fK3Aztwe3TQvX2QKQSMyxDiQJ1dR1P8u/TvWb5bWty2m3KTXlvqMs7Ky2ieVbqVBKSEcHJL4JB3ZwfeLfcQRnTpY7mT7PLZiOdbJSkillzgA44KMScLsBBAOBkuNxu0/6epcQv9s0+LfJzauxBuVJJDRckdXPJ+YcDJH8QrTrN2sNcsxsk2LZyjd9nXaCWj439VPH3RwcZ/hFaVNGc+gUUUUyAooooAxfFVxZxeG9RS7ltVQ25ytwzbCCQBkJ82MkD5eeah0G7tYLi/sZJrKO4fUbjy4oncM/SQ5D9Ww4J25Xniiis2/eO2FNOhf1/CxmamsEGp2+nzx2CwxajYyWKN9o3KdpX+Ebd2I2287ePm973i3UdMg0W+0y4mtUkNqJPKuBJ5ewuEBYx8gbiBxz+FFFS3ZM1p01OdNN/wBaFfVtU0qHxHplx9qsSkEl2853SvIjxwjdtCZXIX7wbt05q7YJdS6nc6vYxWEtpfi2KS+bKsjQhCSWBBG4bhtAAyCcmiinF3k0RWgqdKMl1VvxZfM2s+VkWFh5nl5x9tfG/djGfK6bec468Y/irN1CeUCeHXbrTItPc3O6GN5PNltxHx0I+YKXLYB42455ooqpaIwo2lO1rE1rZjUYrcCO2Giw/Zp7BYzKrkKu4bh8oAB2EA56HIz0u3uxL+1kbygQpQFt2fmki4GOOuOvfHbNFFPpcTu6nKFpsTU75V8oNJKXIXduOI4hk54zjHTjGO+a0KKKaM59PQxLqNNBMuoQpDFYJEfPQLISp8zcWAXIxh5CcLnOMnHQaFNKkkvtOFoli0k9xqP32Zn24LIFyM7kwRg98c5yUVL3No6xTfV2/IrxyW0vh21kQ2phaexKn97s5aErj+LPTbnj7u7+KujoopxZNZW+9/oQXdpBfWk1rcxiSGVGjdSSMhgQeRyOCRxWOtvbXU0Ol6mIHksJbea0IMoJYISGy3U5ST+JuB83uUUMVJuz121JnaL/AITOBSYPOGnyEA7/ADdvmJnH8G3IHX5s4xxmtmiihdRVFZR9AoooqjI//9k=' -response = requests.get('https://ocr.holey.cc/ncku?base64_str=%s' % image) #.split('base64,')[1]) -print(response.content) \ No newline at end of file +response = requests.get('https://ocr.holey.cc/ncku?base64_str=%s' % image) # .split('base64,')[1]) +print(response.content) diff --git a/unfinished/theb.ai/__init__.py b/unfinished/theb.ai/__init__.py index a90a32f5..e6bcb8c0 100644 --- a/unfinished/theb.ai/__init__.py +++ b/unfinished/theb.ai/__init__.py @@ -1,46 +1,49 @@ -from curl_cffi import requests -from json import loads -from re import findall -from threading import Thread -from queue import Queue, Empty +from json import loads +from queue import Queue, Empty +from re import findall +from threading import Thread + +from curl_cffi import requests + class Completion: # experimental part1 = '{"role":"assistant","id":"chatcmpl' part2 = '"},"index":0,"finish_reason":null}]}}' regex = rf'{part1}(.*){part2}' - - timer = None - message_queue = Queue() + + timer = None + message_queue = Queue() stream_completed = False - + def request(): headers = { - 'authority' : 'chatbot.theb.ai', + 'authority': 'chatbot.theb.ai', 'content-type': 'application/json', - 'origin' : 'https://chatbot.theb.ai', - 'user-agent' : 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36', + 'origin': 'https://chatbot.theb.ai', + 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36', } - requests.post('https://chatbot.theb.ai/api/chat-process', headers=headers, content_callback=Completion.handle_stream_response, - json = { - 'prompt' : 'hello world', - 'options': {} - } - ) + requests.post('https://chatbot.theb.ai/api/chat-process', headers=headers, + content_callback=Completion.handle_stream_response, + json={ + 'prompt': 'hello world', + 'options': {} + } + ) Completion.stream_completed = True @staticmethod def create(): Thread(target=Completion.request).start() - + while Completion.stream_completed != True or not Completion.message_queue.empty(): try: message = Completion.message_queue.get(timeout=0.01) for message in findall(Completion.regex, message): yield loads(Completion.part1 + message + Completion.part2) - + except Empty: pass @@ -48,10 +51,12 @@ class Completion: def handle_stream_response(response): Completion.message_queue.put(response.decode()) + def start(): for message in Completion.create(): yield message['delta'] + if __name__ == '__main__': for message in start(): print(message) diff --git a/unfinished/vercelai/v2.py b/unfinished/vercelai/v2.py index d1a4ad05..176ee342 100644 --- a/unfinished/vercelai/v2.py +++ b/unfinished/vercelai/v2.py @@ -1,6 +1,5 @@ import requests - token = requests.get('https://play.vercel.ai/openai.jpeg', headers={ 'authority': 'play.vercel.ai', 'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3', @@ -15,7 +14,7 @@ headers = { 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36' } -for chunk in requests.post('https://play.vercel.ai/api/generate', headers=headers, stream = True, json = { +for chunk in requests.post('https://play.vercel.ai/api/generate', headers=headers, stream=True, json={ 'prompt': 'hi', 'model': 'openai:gpt-3.5-turbo', 'temperature': 0.7, @@ -25,5 +24,4 @@ for chunk in requests.post('https://play.vercel.ai/api/generate', headers=header 'frequencyPenalty': 1, 'presencePenalty': 1, 'stopSequences': []}).iter_lines(): - - print(chunk) \ No newline at end of file + print(chunk) diff --git a/unfinished/writesonic/__init__.py b/unfinished/writesonic/__init__.py index 7df6f393..ce684912 100644 --- a/unfinished/writesonic/__init__.py +++ b/unfinished/writesonic/__init__.py @@ -1,29 +1,33 @@ -from requests import Session -from names import get_first_name, get_last_name -from random import choice -from requests import post -from time import time -from colorama import Fore, init; init() +from random import choice +from time import time + +from colorama import Fore, init; +from names import get_first_name, get_last_name +from requests import Session +from requests import post + +init() + class logger: @staticmethod def info(string) -> print: import datetime now = datetime.datetime.now() - return print(f"{Fore.CYAN}{now.strftime('%Y-%m-%d %H:%M:%S')} {Fore.BLUE}INFO {Fore.MAGENTA}__main__ -> {Fore.RESET}{string}") + return print( + f"{Fore.CYAN}{now.strftime('%Y-%m-%d %H:%M:%S')} {Fore.BLUE}INFO {Fore.MAGENTA}__main__ -> {Fore.RESET}{string}") + class SonicResponse: - class Completion: - class Choices: def __init__(self, choice: dict) -> None: - self.text = choice['text'] - self.content = self.text.encode() - self.index = choice['index'] - self.logprobs = choice['logprobs'] - self.finish_reason = choice['finish_reason'] - + self.text = choice['text'] + self.content = self.text.encode() + self.index = choice['index'] + self.logprobs = choice['logprobs'] + self.finish_reason = choice['finish_reason'] + def __repr__(self) -> str: return f'''<__main__.APIResponse.Completion.Choices(\n text = {self.text.encode()},\n index = {self.index},\n logprobs = {self.logprobs},\n finish_reason = {self.finish_reason})object at 0x1337>''' @@ -32,127 +36,128 @@ class SonicResponse: class Usage: def __init__(self, usage_dict: dict) -> None: - self.prompt_tokens = usage_dict['prompt_chars'] - self.completion_tokens = usage_dict['completion_chars'] - self.total_tokens = usage_dict['total_chars'] + self.prompt_tokens = usage_dict['prompt_chars'] + self.completion_tokens = usage_dict['completion_chars'] + self.total_tokens = usage_dict['total_chars'] def __repr__(self): return f'''<__main__.APIResponse.Usage(\n prompt_tokens = {self.prompt_tokens},\n completion_tokens = {self.completion_tokens},\n total_tokens = {self.total_tokens})object at 0x1337>''' - + def __init__(self, response_dict: dict) -> None: - - self.response_dict = response_dict - self.id = response_dict['id'] - self.object = response_dict['object'] - self.created = response_dict['created'] - self.model = response_dict['model'] - self.completion = self.Completion(response_dict['choices']) - self.usage = self.Usage(response_dict['usage']) + self.response_dict = response_dict + self.id = response_dict['id'] + self.object = response_dict['object'] + self.created = response_dict['created'] + self.model = response_dict['model'] + self.completion = self.Completion(response_dict['choices']) + self.usage = self.Usage(response_dict['usage']) def json(self) -> dict: return self.response_dict - + + class Account: session = Session() session.headers = { - "connection" : "keep-alive", - "sec-ch-ua" : "\"Not_A Brand\";v=\"99\", \"Google Chrome\";v=\"109\", \"Chromium\";v=\"109\"", - "accept" : "application/json, text/plain, */*", - "content-type" : "application/json", - "sec-ch-ua-mobile" : "?0", - "user-agent" : "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/109.0.0.0 Safari/537.36", + "connection": "keep-alive", + "sec-ch-ua": "\"Not_A Brand\";v=\"99\", \"Google Chrome\";v=\"109\", \"Chromium\";v=\"109\"", + "accept": "application/json, text/plain, */*", + "content-type": "application/json", + "sec-ch-ua-mobile": "?0", + "user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/109.0.0.0 Safari/537.36", "sec-ch-ua-platform": "\"Windows\"", - "sec-fetch-site" : "same-origin", - "sec-fetch-mode" : "cors", - "sec-fetch-dest" : "empty", + "sec-fetch-site": "same-origin", + "sec-fetch-mode": "cors", + "sec-fetch-dest": "empty", # "accept-encoding" : "gzip, deflate, br", - "accept-language" : "en-GB,en-US;q=0.9,en;q=0.8", - "cookie" : "" + "accept-language": "en-GB,en-US;q=0.9,en;q=0.8", + "cookie": "" } - + @staticmethod def get_user(): password = f'0opsYouGoTme@1234' - f_name = get_first_name() - l_name = get_last_name() - hosts = ['gmail.com', 'protonmail.com', 'proton.me', 'outlook.com'] - + f_name = get_first_name() + l_name = get_last_name() + hosts = ['gmail.com', 'protonmail.com', 'proton.me', 'outlook.com'] + return { - "email" : f"{f_name.lower()}.{l_name.lower()}@{choice(hosts)}", - "password" : password, - "confirm_password" : password, - "full_name" : f'{f_name} {l_name}' + "email": f"{f_name.lower()}.{l_name.lower()}@{choice(hosts)}", + "password": password, + "confirm_password": password, + "full_name": f'{f_name} {l_name}' } @staticmethod def create(logging: bool = False): while True: try: - user = Account.get_user() - start = time() - response = Account.session.post("https://app.writesonic.com/api/session-login", json = user | { - "utmParams" : "{}", - "visitorId" : "0", - "locale" : "en", - "userAgent" : "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/109.0.0.0 Safari/537.36", - "signInWith" : "password", - "request_type" : "signup", + user = Account.get_user() + start = time() + response = Account.session.post("https://app.writesonic.com/api/session-login", json=user | { + "utmParams": "{}", + "visitorId": "0", + "locale": "en", + "userAgent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/109.0.0.0 Safari/537.36", + "signInWith": "password", + "request_type": "signup", }) - + if logging: logger.info(f"\x1b[31mregister success\x1b[0m : '{response.text[:30]}...' ({int(time() - start)}s)") logger.info(f"\x1b[31mid\x1b[0m : '{response.json()['id']}'") logger.info(f"\x1b[31mtoken\x1b[0m : '{response.json()['token'][:30]}...'") - + start = time() - response = Account.session.post("https://api.writesonic.com/v1/business/set-business-active", headers={"authorization": "Bearer " + response.json()['token']}) + response = Account.session.post("https://api.writesonic.com/v1/business/set-business-active", + headers={"authorization": "Bearer " + response.json()['token']}) key = response.json()["business"]["api_key"] if logging: logger.info(f"\x1b[31mgot key\x1b[0m : '{key}' ({int(time() - start)}s)") return Account.AccountResponse(user['email'], user['password'], key) - + except Exception as e: if logging: logger.info(f"\x1b[31merror\x1b[0m : '{e}'") continue - + class AccountResponse: def __init__(self, email, password, key): - self.email = email + self.email = email self.password = password - self.key = key - + self.key = key + class Completion: def create( - api_key: str, - prompt: str, - enable_memory: bool = False, - enable_google_results: bool = False, - history_data: list = []) -> SonicResponse: - - response = post('https://api.writesonic.com/v2/business/content/chatsonic?engine=premium', headers = {"X-API-KEY": api_key}, - json = { - "enable_memory" : enable_memory, - "enable_google_results" : enable_google_results, - "input_text" : prompt, - "history_data" : history_data}).json() + api_key: str, + prompt: str, + enable_memory: bool = False, + enable_google_results: bool = False, + history_data: list = []) -> SonicResponse: + response = post('https://api.writesonic.com/v2/business/content/chatsonic?engine=premium', + headers={"X-API-KEY": api_key}, + json={ + "enable_memory": enable_memory, + "enable_google_results": enable_google_results, + "input_text": prompt, + "history_data": history_data}).json() return SonicResponse({ - 'id' : f'cmpl-premium-{int(time())}', - 'object' : 'text_completion', - 'created': int(time()), - 'model' : 'premium', - - 'choices': [{ - 'text' : response['message'], - 'index' : 0, - 'logprobs' : None, - 'finish_reason' : 'stop' - }], - - 'usage': { - 'prompt_chars' : len(prompt), - 'completion_chars' : len(response['message']), - 'total_chars' : len(prompt) + len(response['message']) - } - }) \ No newline at end of file + 'id': f'cmpl-premium-{int(time())}', + 'object': 'text_completion', + 'created': int(time()), + 'model': 'premium', + + 'choices': [{ + 'text': response['message'], + 'index': 0, + 'logprobs': None, + 'finish_reason': 'stop' + }], + + 'usage': { + 'prompt_chars': len(prompt), + 'completion_chars': len(response['message']), + 'total_chars': len(prompt) + len(response['message']) + } + }) -- cgit v1.2.3