From 348670fe35369298af8e289af9d523beecfed4c3 Mon Sep 17 00:00:00 2001 From: Heiner Lohaus Date: Tue, 26 Sep 2023 00:52:29 +0200 Subject: "create_async" support for BaseProvider, by using ThreadPoolExecutor Default Model for Vercel --- g4f/Provider/AItianhuSpace.py | 73 +++++++++++++++++++++++++++++++++++++++++++ g4f/Provider/Aivvm.py | 5 +-- g4f/Provider/__init__.py | 2 ++ 3 files changed, 78 insertions(+), 2 deletions(-) create mode 100644 g4f/Provider/AItianhuSpace.py (limited to 'g4f/Provider') diff --git a/g4f/Provider/AItianhuSpace.py b/g4f/Provider/AItianhuSpace.py new file mode 100644 index 00000000..8beb3355 --- /dev/null +++ b/g4f/Provider/AItianhuSpace.py @@ -0,0 +1,73 @@ +from __future__ import annotations + +import random, json + +from g4f.requests import AsyncSession, StreamRequest +from .base_provider import AsyncGeneratorProvider, format_prompt + +domains = { + "gpt-3.5-turbo": ".aitianhu.space", + "gpt-4": ".aitianhu.website", +} + +class AItianhuSpace(AsyncGeneratorProvider): + url = "https://chat3.aiyunos.top/" + working = True + supports_gpt_35_turbo = True + + @classmethod + async def create_async_generator( + cls, + model: str, + messages: list[dict[str, str]], + stream: bool = True, + **kwargs + ) -> str: + if not model: + model = "gpt-3.5-turbo" + elif not model in domains: + raise ValueError(f"Model are not supported: {model}") + + chars = 'abcdefghijklmnopqrstuvwxyz0123456789' + rand = ''.join(random.choice(chars) for _ in range(6)) + domain = domains[model] + url = f'https://{rand}{domain}/api/chat-process' + + headers = { + "User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36", + } + async with AsyncSession(headers=headers, impersonate="chrome107", verify=False) as session: + data = { + "prompt": format_prompt(messages), + "options": {}, + "systemMessage": "You are ChatGPT, a large language model trained by OpenAI. Follow the user's instructions carefully.", + "temperature": 0.8, + "top_p": 1, + **kwargs + } + async with StreamRequest(session, "POST", url, json=data) as response: + response.raise_for_status() + async for line in response.content: + line = json.loads(line.rstrip()) + if "detail" in line: + content = line["detail"]["choices"][0]["delta"].get("content") + if content: + yield content + elif "message" in line and "AI-4接口非常昂贵" in line["message"]: + raise RuntimeError("Rate limit for GPT 4 reached") + else: + raise RuntimeError("Response: {line}") + + + @classmethod + @property + def params(cls): + params = [ + ("model", "str"), + ("messages", "list[dict[str, str]]"), + ("stream", "bool"), + ("temperature", "float"), + ("top_p", "int"), + ] + param = ", ".join([": ".join(p) for p in params]) + return f"g4f.provider.{cls.__name__} supports: ({param})" diff --git a/g4f/Provider/Aivvm.py b/g4f/Provider/Aivvm.py index c38c4a74..7a3d57bd 100644 --- a/g4f/Provider/Aivvm.py +++ b/g4f/Provider/Aivvm.py @@ -60,9 +60,10 @@ class Aivvm(BaseProvider): response = requests.post( "https://chat.aivvm.com/api/chat", headers=headers, json=json_data, stream=True) + response.raise_for_status() - for line in response.iter_content(chunk_size=1048): - yield line.decode('utf-8') + for chunk in response.iter_content(chunk_size=None): + yield chunk.decode('utf-8') @classmethod @property diff --git a/g4f/Provider/__init__.py b/g4f/Provider/__init__.py index aa19ade3..ebe01603 100644 --- a/g4f/Provider/__init__.py +++ b/g4f/Provider/__init__.py @@ -4,6 +4,7 @@ from .Aichat import Aichat from .Ails import Ails from .AiService import AiService from .AItianhu import AItianhu +from .AItianhuSpace import AItianhuSpace from .Aivvm import Aivvm from .Bard import Bard from .Bing import Bing @@ -52,6 +53,7 @@ __all__ = [ 'Ails', 'AiService', 'AItianhu', + 'AItianhuSpace', 'Aivvm', 'Bard', 'Bing', -- cgit v1.2.3