diff options
author | H Lohaus <hlohaus@users.noreply.github.com> | 2024-04-06 21:01:27 +0200 |
---|---|---|
committer | GitHub <noreply@github.com> | 2024-04-06 21:01:27 +0200 |
commit | 6e3f350f52d1cef215d7d9ac3100596c456a5d70 (patch) | |
tree | 7682dfe5fdc3a4e6ed17fbdd9dc1c6bf274c2277 /g4f | |
parent | Update types.py (diff) | |
download | gpt4free-6e3f350f52d1cef215d7d9ac3100596c456a5d70.tar gpt4free-6e3f350f52d1cef215d7d9ac3100596c456a5d70.tar.gz gpt4free-6e3f350f52d1cef215d7d9ac3100596c456a5d70.tar.bz2 gpt4free-6e3f350f52d1cef215d7d9ac3100596c456a5d70.tar.lz gpt4free-6e3f350f52d1cef215d7d9ac3100596c456a5d70.tar.xz gpt4free-6e3f350f52d1cef215d7d9ac3100596c456a5d70.tar.zst gpt4free-6e3f350f52d1cef215d7d9ac3100596c456a5d70.zip |
Diffstat (limited to '')
-rw-r--r-- | g4f/client/async.py | 48 |
1 files changed, 18 insertions, 30 deletions
diff --git a/g4f/client/async.py b/g4f/client/async.py index 6b23a9d5..bcd1a502 100644 --- a/g4f/client/async.py +++ b/g4f/client/async.py @@ -6,49 +6,38 @@ import time import random import string +from .types import BaseProvider, ProviderType, FinishReason from .stubs import ChatCompletion, ChatCompletionChunk, Image, ImagesResponse -from .typing import Union, Iterator, Messages, ImageType -from .providers.types import BaseProvider, ProviderType, FinishReason -from .image import ImageResponse as ImageProviderResponse -from .errors import NoImageResponseError, RateLimitError, MissingAuthError -from . import get_model_and_provider, get_last_provider +from ..typing import Union, Iterator, Messages, ImageType, AsyncIerator +from ..image import ImageResponse as ImageProviderResponse +from ..errors import NoImageResponseError, RateLimitError, MissingAuthError +from .. import get_model_and_provider, get_last_provider +from .helper import read_json from .Provider.BingCreateImages import BingCreateImages from .Provider.needs_auth import Gemini, OpenaiChat -from .Provider.You import You -from .helper import read_json +from ..Provider.You import You -def iter_response( - response: iter[str], +async def iter_response( + response: AsyncIerator[str], stream: bool, response_format: dict = None, max_tokens: int = None, stop: list = None -) -> IterResponse: +) -> AsyncIterResponse: content = "" finish_reason = None completion_id = ''.join(random.choices(string.ascii_letters + string.digits, k=28)) - for idx, chunk in enumerate(response): + count: int = 0 + async for idx, chunk in response: if isinstance(chunk, FinishReason): finish_reason = chunk.reason break content += str(chunk) - if max_tokens is not None and idx + 1 >= max_tokens: + count += 1 + if max_tokens is not None and count >= max_tokens: finish_reason = "length" - first = -1 - word = None - if stop is not None: - for word in list(stop): - first = content.find(word) - if first != -1: - content = content[:first] - break - if stream and first != -1: - first = chunk.find(word) - if first != -1: - chunk = chunk[:first] - else: - first = 0 + first, content, chunk = find_stop(stop, content, chunk) if first != -1: finish_reason = "stop" if stream: @@ -64,16 +53,15 @@ def iter_response( content = read_json(content) yield ChatCompletion(content, finish_reason, completion_id, int(time.time())) -def iter_append_model_and_provider(response: IterResponse) -> IterResponse: +async def iter_append_model_and_provider(response: AsyncIterResponse) -> IterResponse: last_provider = None - for chunk in response: + async for chunk in response: last_provider = get_last_provider(True) if last_provider is None else last_provider chunk.model = last_provider.get("model") chunk.provider = last_provider.get("name") yield chunk class Client(): - def __init__( self, api_key: str = None, @@ -222,4 +210,4 @@ class Images(): result = ImagesResponse([Image(image)for image in result]) if result is None: raise NoImageResponseError() - return result
\ No newline at end of file + return result |