summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorabc <98614666+xtekky@users.noreply.github.com>2024-05-15 00:45:13 +0200
committerabc <98614666+xtekky@users.noreply.github.com>2024-05-15 00:45:13 +0200
commit1399758a4e000b0fe1cd3d9e86d1a4807d9e41e8 (patch)
treeaabd2aa2013c7526c5e4de0a9dc2e05da81a5a4c
parent. (diff)
downloadgpt4free-1399758a4e000b0fe1cd3d9e86d1a4807d9e41e8.tar
gpt4free-1399758a4e000b0fe1cd3d9e86d1a4807d9e41e8.tar.gz
gpt4free-1399758a4e000b0fe1cd3d9e86d1a4807d9e41e8.tar.bz2
gpt4free-1399758a4e000b0fe1cd3d9e86d1a4807d9e41e8.tar.lz
gpt4free-1399758a4e000b0fe1cd3d9e86d1a4807d9e41e8.tar.xz
gpt4free-1399758a4e000b0fe1cd3d9e86d1a4807d9e41e8.tar.zst
gpt4free-1399758a4e000b0fe1cd3d9e86d1a4807d9e41e8.zip
-rw-r--r--README.md2
-rw-r--r--g4f/Provider/You.py20
-rw-r--r--g4f/Provider/you/har_file.py34
-rw-r--r--g4f/models.py9
4 files changed, 51 insertions, 14 deletions
diff --git a/README.md b/README.md
index b5c97854..2f3d19a7 100644
--- a/README.md
+++ b/README.md
@@ -301,7 +301,7 @@ set G4F_PROXY=http://host:port
While we wait for gpt-5, here is a list of new models that are at least better than gpt-3.5-turbo. **Some are better than gpt-4**. Expect this list to grow.
| Website | Provider | parameters | better than |
-| ------ | ------- | ------ | ------ |
+| ------ | ------- | ------ | ------ |
| [claude-3-opus](https://anthropic.com/) | `g4f.Provider.You` | ?B | gpt-4-0125-preview |
| [command-r+](https://txt.cohere.com/command-r-plus-microsoft-azure/) | `g4f.Provider.HuggingChat` | 104B | gpt-4-0314 |
| [llama-3-70b](https://meta.ai/) | `g4f.Provider.Llama` or `DeepInfra` | 70B | gpt-4-0314 |
diff --git a/g4f/Provider/You.py b/g4f/Provider/You.py
index e7958f71..b4892035 100644
--- a/g4f/Provider/You.py
+++ b/g4f/Provider/You.py
@@ -1,5 +1,3 @@
-from __future__ import annotations
-
import re
import json
import base64
@@ -24,6 +22,7 @@ class You(AsyncGeneratorProvider, ProviderModelMixin):
image_models = ["dall-e"]
models = [
default_model,
+ "gpt-4o",
"gpt-4",
"gpt-4-turbo",
"claude-instant",
@@ -42,7 +41,8 @@ class You(AsyncGeneratorProvider, ProviderModelMixin):
*image_models
]
model_aliases = {
- "claude-v2": "claude-2"
+ "claude-v2": "claude-2",
+ "gpt-4o": "gpt-4o",
}
_cookies = None
_cookies_used = 0
@@ -99,7 +99,9 @@ class You(AsyncGeneratorProvider, ProviderModelMixin):
"selectedChatMode": chat_mode,
}
if chat_mode == "custom":
+ # print(f"You model: {model}")
params["selectedAIModel"] = model.replace("-", "_")
+
async with (session.post if chat_mode == "default" else session.get)(
f"{cls.url}/api/streamingSearch",
data=data,
@@ -183,7 +185,15 @@ class You(AsyncGeneratorProvider, ProviderModelMixin):
@classmethod
async def create_cookies(cls, client: StreamSession) -> Cookies:
if not cls._telemetry_ids:
- cls._telemetry_ids = await get_telemetry_ids()
+ try:
+ cls._telemetry_ids = await get_telemetry_ids()
+ except RuntimeError as e:
+ if str(e) == "Event loop is closed":
+ if debug.logging:
+ print("Event loop is closed error occurred in create_cookies.")
+ else:
+ raise
+
user_uuid = str(uuid.uuid4())
telemetry_id = cls._telemetry_ids.pop()
if debug.logging:
@@ -212,4 +222,4 @@ class You(AsyncGeneratorProvider, ProviderModelMixin):
'stytch_session_jwt': session["session_jwt"],
'ydc_stytch_session': session["session_token"],
'ydc_stytch_session_jwt': session["session_jwt"],
- }
+ } \ No newline at end of file
diff --git a/g4f/Provider/you/har_file.py b/g4f/Provider/you/har_file.py
index 8089d686..cfdca12f 100644
--- a/g4f/Provider/you/har_file.py
+++ b/g4f/Provider/you/har_file.py
@@ -4,11 +4,14 @@ import json
import os
import os.path
import random
+import logging
from ...requests import StreamSession, raise_for_status
from ...errors import MissingRequirementsError
from ... import debug
+logging.basicConfig(level=logging.ERROR)
+
class NoValidHarFileError(Exception):
...
@@ -62,10 +65,16 @@ def parseHAREntry(entry) -> arkReq:
return tmpArk
async def sendRequest(tmpArk: arkReq, proxy: str = None):
- async with StreamSession(headers=tmpArk.arkHeaders, cookies=tmpArk.arkCookies, proxy=proxy) as session:
- async with session.post(tmpArk.arkURL, data=tmpArk.arkBody) as response:
- await raise_for_status(response)
- return await response.text()
+ try:
+ async with StreamSession(headers=tmpArk.arkHeaders, cookies=tmpArk.arkCookies, proxy=proxy) as session:
+ async with session.post(tmpArk.arkURL, data=tmpArk.arkBody) as response:
+ await raise_for_status(response)
+ return await response.text()
+ except RuntimeError as e:
+ if str(e) == "Event loop is closed":
+ print("Event loop is closed error occurred in sendRequest.")
+ else:
+ raise
async def create_telemetry_id(proxy: str = None):
global chatArks
@@ -78,9 +87,9 @@ async def get_telemetry_ids(proxy: str = None) -> list:
return [await create_telemetry_id(proxy)]
except NoValidHarFileError as e:
if debug.logging:
- print(e)
+ logging.error(e)
if debug.logging:
- print('Getting telemetry_id for you.com with nodriver')
+ logging.error('Getting telemetry_id for you.com with nodriver')
try:
from nodriver import start
except ImportError:
@@ -100,6 +109,15 @@ async def get_telemetry_ids(proxy: str = None) -> list:
)
return [await get_telemetry_id()]
+
finally:
- if page is not None:
- await page.close() \ No newline at end of file
+ try:
+ if page is not None:
+ await page.close()
+
+ if browser is not None:
+ await browser.close()
+
+ except Exception as e:
+ if debug.logging:
+ logging.error(e)
diff --git a/g4f/models.py b/g4f/models.py
index e3da0363..0c5eaa9f 100644
--- a/g4f/models.py
+++ b/g4f/models.py
@@ -100,6 +100,14 @@ gpt_4 = Model(
])
)
+gpt_4o = Model(
+ name = 'gpt-4o',
+ base_provider = 'openai',
+ best_provider = RetryProvider([
+ You
+ ])
+)
+
gpt_4_turbo = Model(
name = 'gpt-4-turbo',
base_provider = 'openai',
@@ -330,6 +338,7 @@ class ModelUtils:
'gpt-3.5-long': gpt_35_long,
# gpt-4
+ 'gpt-4o' : gpt_4o,
'gpt-4' : gpt_4,
'gpt-4-0613' : gpt_4_0613,
'gpt-4-32k' : gpt_4_32k,