diff options
author | kqlio67 <kqlio67@users.noreply.github.com> | 2024-09-05 08:58:28 +0200 |
---|---|---|
committer | kqlio67 <kqlio67@users.noreply.github.com> | 2024-09-05 08:58:28 +0200 |
commit | f2f04a00b12abf49b74262ade643942e5d1571ea (patch) | |
tree | 2c71cfa978db5b3a833765a6487335b029147d6d | |
parent | Updated g4f/gui/server/api.py (diff) | |
download | gpt4free-f2f04a00b12abf49b74262ade643942e5d1571ea.tar gpt4free-f2f04a00b12abf49b74262ade643942e5d1571ea.tar.gz gpt4free-f2f04a00b12abf49b74262ade643942e5d1571ea.tar.bz2 gpt4free-f2f04a00b12abf49b74262ade643942e5d1571ea.tar.lz gpt4free-f2f04a00b12abf49b74262ade643942e5d1571ea.tar.xz gpt4free-f2f04a00b12abf49b74262ade643942e5d1571ea.tar.zst gpt4free-f2f04a00b12abf49b74262ade643942e5d1571ea.zip |
-rw-r--r-- | g4f/Provider/Blackbox.py | 77 |
1 files changed, 62 insertions, 15 deletions
diff --git a/g4f/Provider/Blackbox.py b/g4f/Provider/Blackbox.py index 3c9e394e..9fab4a09 100644 --- a/g4f/Provider/Blackbox.py +++ b/g4f/Provider/Blackbox.py @@ -3,11 +3,12 @@ from __future__ import annotations import uuid import secrets import re -from aiohttp import ClientSession, ClientResponse +import base64 +from aiohttp import ClientSession from typing import AsyncGenerator, Optional from ..typing import AsyncResult, Messages, ImageType -from ..image import to_data_uri +from ..image import to_data_uri, ImageResponse from .base_provider import AsyncGeneratorProvider, ProviderModelMixin class Blackbox(AsyncGeneratorProvider, ProviderModelMixin): @@ -20,12 +21,25 @@ class Blackbox(AsyncGeneratorProvider, ProviderModelMixin): "llama-3.1-8b", 'llama-3.1-70b', 'llama-3.1-405b', + 'ImageGeneration', ] model_aliases = { "gemini-flash": "gemini-1.5-flash", } + agent_mode_map = { + 'ImageGeneration': {"mode": True, "id": "ImageGenerationLV45LJp", "name": "Image Generation"}, + } + + model_id_map = { + "blackbox": {}, + "gemini-1.5-flash": {'mode': True, 'id': 'Gemini'}, + "llama-3.1-8b": {'mode': True, 'id': "llama-3.1-8b"}, + 'llama-3.1-70b': {'mode': True, 'id': "llama-3.1-70b"}, + 'llama-3.1-405b': {'mode': True, 'id': "llama-3.1-405b"} + } + @classmethod def get_model(cls, model: str) -> str: if model in cls.models: @@ -36,6 +50,15 @@ class Blackbox(AsyncGeneratorProvider, ProviderModelMixin): return cls.default_model @classmethod + async def download_image_to_base64_url(cls, url: str) -> str: + async with ClientSession() as session: + async with session.get(url) as response: + image_data = await response.read() + base64_data = base64.b64encode(image_data).decode('utf-8') + mime_type = response.headers.get('Content-Type', 'image/jpeg') + return f"data:{mime_type};base64,{base64_data}" + + @classmethod async def create_async_generator( cls, model: str, @@ -44,7 +67,7 @@ class Blackbox(AsyncGeneratorProvider, ProviderModelMixin): image: Optional[ImageType] = None, image_name: Optional[str] = None, **kwargs - ) -> AsyncGenerator[str, None]: + ) -> AsyncGenerator[AsyncResult, None]: if image is not None: messages[-1]["data"] = { "fileText": image_name, @@ -72,20 +95,12 @@ class Blackbox(AsyncGeneratorProvider, ProviderModelMixin): model = cls.get_model(model) # Resolve the model alias - model_id_map = { - "blackbox": {}, - "gemini-1.5-flash": {'mode': True, 'id': 'Gemini'}, - "llama-3.1-8b": {'mode': True, 'id': "llama-3.1-8b"}, - 'llama-3.1-70b': {'mode': True, 'id': "llama-3.1-70b"}, - 'llama-3.1-405b': {'mode': True, 'id': "llama-3.1-405b"} - } - data = { "messages": messages, "id": random_id, "userId": random_user_id, "codeModelMode": True, - "agentMode": {}, + "agentMode": cls.agent_mode_map.get(model, {}), "trendingAgentMode": {}, "isMicMode": False, "isChromeExt": False, @@ -93,7 +108,7 @@ class Blackbox(AsyncGeneratorProvider, ProviderModelMixin): "webSearchMode": False, "userSystemPrompt": "", "githubToken": None, - "trendingAgentModel": model_id_map.get(model, {}), # Default to empty dict if model not found + "trendingAgentModel": cls.model_id_map.get(model, {}), "maxTokens": None } @@ -101,9 +116,41 @@ class Blackbox(AsyncGeneratorProvider, ProviderModelMixin): f"{cls.url}/api/chat", json=data, proxy=proxy ) as response: response.raise_for_status() + full_response = "" + buffer = "" + image_base64_url = None async for chunk in response.content.iter_any(): if chunk: - # Decode the chunk and clean up unwanted prefixes using a regex decoded_chunk = chunk.decode() cleaned_chunk = re.sub(r'\$@\$.+?\$@\$|\$@\$', '', decoded_chunk) - yield cleaned_chunk + + buffer += cleaned_chunk + + # Check if there's a complete image line in the buffer + image_match = re.search(r'!\[Generated Image\]\((https?://[^\s\)]+)\)', buffer) + if image_match: + image_url = image_match.group(1) + # Download the image and convert to base64 URL + image_base64_url = await cls.download_image_to_base64_url(image_url) + + # Remove the image line from the buffer + buffer = re.sub(r'!\[Generated Image\]\(https?://[^\s\)]+\)', '', buffer) + + # Send text line by line + lines = buffer.split('\n') + for line in lines[:-1]: + if line.strip(): + full_response += line + '\n' + yield line + '\n' + buffer = lines[-1] # Keep the last incomplete line in the buffer + + # Send the remaining buffer if it's not empty + if buffer.strip(): + full_response += buffer + yield buffer + + # If an image was found, send it as ImageResponse + if image_base64_url: + alt_text = "Generated Image" + image_response = ImageResponse(image_base64_url, alt=alt_text) + yield image_response |