summaryrefslogtreecommitdiffstats
path: root/g4f/Provider/OpenaiChat.py
diff options
context:
space:
mode:
authorCommenter123321 <36051603+Commenter123321@users.noreply.github.com>2023-10-09 18:02:06 +0200
committerCommenter123321 <36051603+Commenter123321@users.noreply.github.com>2023-10-09 18:02:06 +0200
commit119817c96349807efaf87ee432ce46446542b66a (patch)
tree1dbdf4d4dbf4f6c8a8247274ef500a2f1de765d1 /g4f/Provider/OpenaiChat.py
parentaivvm's no life creator keeps patching it, but I'm just better 😉 (diff)
parentMerge branch 'main' of https://github.com/xtekky/gpt4free (diff)
downloadgpt4free-119817c96349807efaf87ee432ce46446542b66a.tar
gpt4free-119817c96349807efaf87ee432ce46446542b66a.tar.gz
gpt4free-119817c96349807efaf87ee432ce46446542b66a.tar.bz2
gpt4free-119817c96349807efaf87ee432ce46446542b66a.tar.lz
gpt4free-119817c96349807efaf87ee432ce46446542b66a.tar.xz
gpt4free-119817c96349807efaf87ee432ce46446542b66a.tar.zst
gpt4free-119817c96349807efaf87ee432ce46446542b66a.zip
Diffstat (limited to 'g4f/Provider/OpenaiChat.py')
-rw-r--r--g4f/Provider/OpenaiChat.py88
1 files changed, 0 insertions, 88 deletions
diff --git a/g4f/Provider/OpenaiChat.py b/g4f/Provider/OpenaiChat.py
deleted file mode 100644
index f7dc8298..00000000
--- a/g4f/Provider/OpenaiChat.py
+++ /dev/null
@@ -1,88 +0,0 @@
-from __future__ import annotations
-
-from curl_cffi.requests import AsyncSession
-import uuid
-import json
-
-from .base_provider import AsyncProvider, get_cookies, format_prompt
-from ..typing import AsyncGenerator
-
-
-class OpenaiChat(AsyncProvider):
- url = "https://chat.openai.com"
- needs_auth = True
- working = True
- supports_gpt_35_turbo = True
- _access_token = None
-
- @classmethod
- async def create_async(
- cls,
- model: str,
- messages: list[dict[str, str]],
- proxy: str = None,
- access_token: str = None,
- cookies: dict = None,
- **kwargs: dict
- ) -> AsyncGenerator:
- proxies = {"https": proxy}
- if not access_token:
- access_token = await cls.get_access_token(cookies, proxies)
- headers = {
- "Accept": "text/event-stream",
- "Authorization": f"Bearer {access_token}",
- }
- async with AsyncSession(proxies=proxies, headers=headers, impersonate="chrome107") as session:
- messages = [
- {
- "id": str(uuid.uuid4()),
- "author": {"role": "user"},
- "content": {"content_type": "text", "parts": [format_prompt(messages)]},
- },
- ]
- data = {
- "action": "next",
- "messages": messages,
- "conversation_id": None,
- "parent_message_id": str(uuid.uuid4()),
- "model": "text-davinci-002-render-sha",
- "history_and_training_disabled": True,
- }
- response = await session.post("https://chat.openai.com/backend-api/conversation", json=data)
- response.raise_for_status()
- last_message = None
- for line in response.content.decode().splitlines():
- if line.startswith("data: "):
- line = line[6:]
- if line == "[DONE]":
- break
- line = json.loads(line)
- if "message" in line:
- last_message = line["message"]["content"]["parts"][0]
- return last_message
-
-
- @classmethod
- async def get_access_token(cls, cookies: dict = None, proxies: dict = None) -> str:
- if not cls._access_token:
- cookies = cookies if cookies else get_cookies("chat.openai.com")
- async with AsyncSession(proxies=proxies, cookies=cookies, impersonate="chrome107") as session:
- response = await session.get("https://chat.openai.com/api/auth/session")
- response.raise_for_status()
- cls._access_token = response.json()["accessToken"]
- return cls._access_token
-
-
- @classmethod
- @property
- def params(cls):
- params = [
- ("model", "str"),
- ("messages", "list[dict[str, str]]"),
- ("stream", "bool"),
- ("proxy", "str"),
- ("access_token", "str"),
- ("cookies", "dict[str, str]")
- ]
- param = ", ".join([": ".join(p) for p in params])
- return f"g4f.provider.{cls.__name__} supports: ({param})" \ No newline at end of file