summaryrefslogtreecommitdiffstats
path: root/g4f/Provider/nexra/NexraLlama.py
diff options
context:
space:
mode:
authorTekky <98614666+xtekky@users.noreply.github.com>2024-10-15 11:51:53 +0200
committerGitHub <noreply@github.com>2024-10-15 11:51:53 +0200
commit5ed3467d07181e876d957984c16782d687abd3b5 (patch)
tree23bd0fd3481d81fca70ac3c7842cb7ffa8f6497f /g4f/Provider/nexra/NexraLlama.py
parentMerge pull request #2268 from yjg30737/patch-1 (diff)
parentUpdated(docs/client.md) (diff)
downloadgpt4free-5ed3467d07181e876d957984c16782d687abd3b5.tar
gpt4free-5ed3467d07181e876d957984c16782d687abd3b5.tar.gz
gpt4free-5ed3467d07181e876d957984c16782d687abd3b5.tar.bz2
gpt4free-5ed3467d07181e876d957984c16782d687abd3b5.tar.lz
gpt4free-5ed3467d07181e876d957984c16782d687abd3b5.tar.xz
gpt4free-5ed3467d07181e876d957984c16782d687abd3b5.tar.zst
gpt4free-5ed3467d07181e876d957984c16782d687abd3b5.zip
Diffstat (limited to 'g4f/Provider/nexra/NexraLlama.py')
-rw-r--r--g4f/Provider/nexra/NexraLlama.py52
1 files changed, 0 insertions, 52 deletions
diff --git a/g4f/Provider/nexra/NexraLlama.py b/g4f/Provider/nexra/NexraLlama.py
deleted file mode 100644
index 9ed892e8..00000000
--- a/g4f/Provider/nexra/NexraLlama.py
+++ /dev/null
@@ -1,52 +0,0 @@
-from __future__ import annotations
-
-import json
-from aiohttp import ClientSession
-
-from ...typing import AsyncResult, Messages
-from ..base_provider import AsyncGeneratorProvider, ProviderModelMixin
-from ..helper import format_prompt
-
-
-class NexraLlama(AsyncGeneratorProvider, ProviderModelMixin):
- label = "Nexra LLaMA 3.1"
- api_endpoint = "https://nexra.aryahcr.cc/api/chat/complements"
- models = ['llama-3.1']
-
- @classmethod
- async def create_async_generator(
- cls,
- model: str,
- messages: Messages,
- proxy: str = None,
- **kwargs
- ) -> AsyncResult:
- headers = {
- "Content-Type": "application/json"
- }
- async with ClientSession(headers=headers) as session:
- data = {
- "messages": [
- {'role': 'assistant', 'content': ''},
- {'role': 'user', 'content': format_prompt(messages)}
- ],
- "markdown": False,
- "stream": True,
- "model": model
- }
- async with session.post(cls.api_endpoint, json=data, proxy=proxy) as response:
- response.raise_for_status()
- full_response = ''
- async for line in response.content:
- if line:
- messages = line.decode('utf-8').split('\x1e')
- for message_str in messages:
- try:
- message = json.loads(message_str)
- if message.get('message'):
- full_response = message['message']
- if message.get('finish'):
- yield full_response.strip()
- return
- except json.JSONDecodeError:
- pass