summaryrefslogtreecommitdiffstats
path: root/g4f/local/__init__.py
diff options
context:
space:
mode:
authorabc <98614666+xtekky@users.noreply.github.com>2024-03-11 19:26:34 +0100
committerabc <98614666+xtekky@users.noreply.github.com>2024-03-11 19:26:34 +0100
commitb7342b1f130aa867eec17d973b0cab00b16a4507 (patch)
treea0936e607ee1701ab3ae68f2dc47bf503c20bff7 /g4f/local/__init__.py
parent~ (diff)
downloadgpt4free-b7342b1f130aa867eec17d973b0cab00b16a4507.tar
gpt4free-b7342b1f130aa867eec17d973b0cab00b16a4507.tar.gz
gpt4free-b7342b1f130aa867eec17d973b0cab00b16a4507.tar.bz2
gpt4free-b7342b1f130aa867eec17d973b0cab00b16a4507.tar.lz
gpt4free-b7342b1f130aa867eec17d973b0cab00b16a4507.tar.xz
gpt4free-b7342b1f130aa867eec17d973b0cab00b16a4507.tar.zst
gpt4free-b7342b1f130aa867eec17d973b0cab00b16a4507.zip
Diffstat (limited to 'g4f/local/__init__.py')
-rw-r--r--g4f/local/__init__.py109
1 files changed, 109 insertions, 0 deletions
diff --git a/g4f/local/__init__.py b/g4f/local/__init__.py
new file mode 100644
index 00000000..626643fc
--- /dev/null
+++ b/g4f/local/__init__.py
@@ -0,0 +1,109 @@
+import random, string, time, re
+
+from ..typing import Union, Iterator, Messages
+from ..stubs import ChatCompletion, ChatCompletionChunk
+from .core.engine import LocalProvider
+from .core.models import models
+
+IterResponse = Iterator[Union[ChatCompletion, ChatCompletionChunk]]
+
+def read_json(text: str) -> dict:
+ match = re.search(r"```(json|)\n(?P<code>[\S\s]+?)\n```", text)
+ if match:
+ return match.group("code")
+ return text
+
+def iter_response(
+ response: Iterator[str],
+ stream: bool,
+ response_format: dict = None,
+ max_tokens: int = None,
+ stop: list = None
+) -> IterResponse:
+
+ content = ""
+ finish_reason = None
+ completion_id = ''.join(random.choices(string.ascii_letters + string.digits, k=28))
+ for idx, chunk in enumerate(response):
+ content += str(chunk)
+ if max_tokens is not None and idx + 1 >= max_tokens:
+ finish_reason = "length"
+ first = -1
+ word = None
+ if stop is not None:
+ for word in list(stop):
+ first = content.find(word)
+ if first != -1:
+ content = content[:first]
+ break
+ if stream and first != -1:
+ first = chunk.find(word)
+ if first != -1:
+ chunk = chunk[:first]
+ else:
+ first = 0
+ if first != -1:
+ finish_reason = "stop"
+ if stream:
+ yield ChatCompletionChunk(chunk, None, completion_id, int(time.time()))
+ if finish_reason is not None:
+ break
+ finish_reason = "stop" if finish_reason is None else finish_reason
+ if stream:
+ yield ChatCompletionChunk(None, finish_reason, completion_id, int(time.time()))
+ else:
+ if response_format is not None and "type" in response_format:
+ if response_format["type"] == "json_object":
+ content = read_json(content)
+ yield ChatCompletion(content, finish_reason, completion_id, int(time.time()))
+
+def filter_none(**kwargs):
+ for key in list(kwargs.keys()):
+ if kwargs[key] is None:
+ del kwargs[key]
+ return kwargs
+
+class LocalClient():
+ def __init__(
+ self,
+ **kwargs
+ ) -> None:
+ self.chat: Chat = Chat(self)
+
+ @staticmethod
+ def list_models():
+ return list(models.keys())
+
+class Completions():
+ def __init__(self, client: LocalClient):
+ self.client: LocalClient = client
+
+ def create(
+ self,
+ messages: Messages,
+ model: str,
+ stream: bool = False,
+ response_format: dict = None,
+ max_tokens: int = None,
+ stop: Union[list[str], str] = None,
+ **kwargs
+ ) -> Union[ChatCompletion, Iterator[ChatCompletionChunk]]:
+
+ stop = [stop] if isinstance(stop, str) else stop
+ response = LocalProvider.create_completion(
+ model, messages, stream,
+ **filter_none(
+ max_tokens=max_tokens,
+ stop=stop,
+ ),
+ **kwargs
+ )
+ response = iter_response(response, stream, response_format, max_tokens, stop)
+ return response if stream else next(response)
+
+class Chat():
+ completions: Completions
+
+ def __init__(self, client: LocalClient):
+ self.completions = Completions(client)
+ \ No newline at end of file