diff options
author | Tekky <98614666+xtekky@users.noreply.github.com> | 2024-09-07 21:37:24 +0200 |
---|---|---|
committer | GitHub <noreply@github.com> | 2024-09-07 21:37:24 +0200 |
commit | 07fa87b4d180259d1da86afb565e14ac3d60d50b (patch) | |
tree | d474a2bf8bd79cf94bfa48cbaca3917659dd19be /g4f/Provider/Blackbox.py | |
parent | Merge pull request #2206 from Parthsadaria/patch-1 (diff) | |
parent | g4f/models.py g4f/Provider/MagickPen.py (diff) | |
download | gpt4free-0.3.2.6.tar gpt4free-0.3.2.6.tar.gz gpt4free-0.3.2.6.tar.bz2 gpt4free-0.3.2.6.tar.lz gpt4free-0.3.2.6.tar.xz gpt4free-0.3.2.6.tar.zst gpt4free-0.3.2.6.zip |
Diffstat (limited to '')
-rw-r--r-- | g4f/Provider/Blackbox.py | 93 |
1 files changed, 79 insertions, 14 deletions
diff --git a/g4f/Provider/Blackbox.py b/g4f/Provider/Blackbox.py index fd84875e..9fab4a09 100644 --- a/g4f/Provider/Blackbox.py +++ b/g4f/Provider/Blackbox.py @@ -3,11 +3,12 @@ from __future__ import annotations import uuid import secrets import re -from aiohttp import ClientSession, ClientResponse +import base64 +from aiohttp import ClientSession from typing import AsyncGenerator, Optional from ..typing import AsyncResult, Messages, ImageType -from ..image import to_data_uri +from ..image import to_data_uri, ImageResponse from .base_provider import AsyncGeneratorProvider, ProviderModelMixin class Blackbox(AsyncGeneratorProvider, ProviderModelMixin): @@ -20,7 +21,43 @@ class Blackbox(AsyncGeneratorProvider, ProviderModelMixin): "llama-3.1-8b", 'llama-3.1-70b', 'llama-3.1-405b', + 'ImageGeneration', ] + + model_aliases = { + "gemini-flash": "gemini-1.5-flash", + } + + agent_mode_map = { + 'ImageGeneration': {"mode": True, "id": "ImageGenerationLV45LJp", "name": "Image Generation"}, + } + + model_id_map = { + "blackbox": {}, + "gemini-1.5-flash": {'mode': True, 'id': 'Gemini'}, + "llama-3.1-8b": {'mode': True, 'id': "llama-3.1-8b"}, + 'llama-3.1-70b': {'mode': True, 'id': "llama-3.1-70b"}, + 'llama-3.1-405b': {'mode': True, 'id': "llama-3.1-405b"} + } + + @classmethod + def get_model(cls, model: str) -> str: + if model in cls.models: + return model + elif model in cls.model_aliases: + return cls.model_aliases[model] + else: + return cls.default_model + + @classmethod + async def download_image_to_base64_url(cls, url: str) -> str: + async with ClientSession() as session: + async with session.get(url) as response: + image_data = await response.read() + base64_data = base64.b64encode(image_data).decode('utf-8') + mime_type = response.headers.get('Content-Type', 'image/jpeg') + return f"data:{mime_type};base64,{base64_data}" + @classmethod async def create_async_generator( cls, @@ -30,7 +67,7 @@ class Blackbox(AsyncGeneratorProvider, ProviderModelMixin): image: Optional[ImageType] = None, image_name: Optional[str] = None, **kwargs - ) -> AsyncGenerator[str, None]: + ) -> AsyncGenerator[AsyncResult, None]: if image is not None: messages[-1]["data"] = { "fileText": image_name, @@ -55,19 +92,15 @@ class Blackbox(AsyncGeneratorProvider, ProviderModelMixin): async with ClientSession(headers=headers) as session: random_id = secrets.token_hex(16) random_user_id = str(uuid.uuid4()) - model_id_map = { - "blackbox": {}, - "gemini-1.5-flash": {'mode': True, 'id': 'Gemini'}, - "llama-3.1-8b": {'mode': True, 'id': "llama-3.1-8b"}, - 'llama-3.1-70b': {'mode': True, 'id': "llama-3.1-70b"}, - 'llama-3.1-405b': {'mode': True, 'id': "llama-3.1-405b"} - } + + model = cls.get_model(model) # Resolve the model alias + data = { "messages": messages, "id": random_id, "userId": random_user_id, "codeModelMode": True, - "agentMode": {}, + "agentMode": cls.agent_mode_map.get(model, {}), "trendingAgentMode": {}, "isMicMode": False, "isChromeExt": False, @@ -75,7 +108,7 @@ class Blackbox(AsyncGeneratorProvider, ProviderModelMixin): "webSearchMode": False, "userSystemPrompt": "", "githubToken": None, - "trendingAgentModel": model_id_map[model], # if you actually test this on the site, just ask each model "yo", weird behavior imo + "trendingAgentModel": cls.model_id_map.get(model, {}), "maxTokens": None } @@ -83,9 +116,41 @@ class Blackbox(AsyncGeneratorProvider, ProviderModelMixin): f"{cls.url}/api/chat", json=data, proxy=proxy ) as response: response.raise_for_status() + full_response = "" + buffer = "" + image_base64_url = None async for chunk in response.content.iter_any(): if chunk: - # Decode the chunk and clean up unwanted prefixes using a regex decoded_chunk = chunk.decode() cleaned_chunk = re.sub(r'\$@\$.+?\$@\$|\$@\$', '', decoded_chunk) - yield cleaned_chunk + + buffer += cleaned_chunk + + # Check if there's a complete image line in the buffer + image_match = re.search(r'!\[Generated Image\]\((https?://[^\s\)]+)\)', buffer) + if image_match: + image_url = image_match.group(1) + # Download the image and convert to base64 URL + image_base64_url = await cls.download_image_to_base64_url(image_url) + + # Remove the image line from the buffer + buffer = re.sub(r'!\[Generated Image\]\(https?://[^\s\)]+\)', '', buffer) + + # Send text line by line + lines = buffer.split('\n') + for line in lines[:-1]: + if line.strip(): + full_response += line + '\n' + yield line + '\n' + buffer = lines[-1] # Keep the last incomplete line in the buffer + + # Send the remaining buffer if it's not empty + if buffer.strip(): + full_response += buffer + yield buffer + + # If an image was found, send it as ImageResponse + if image_base64_url: + alt_text = "Generated Image" + image_response = ImageResponse(image_base64_url, alt=alt_text) + yield image_response |