summaryrefslogtreecommitdiffstats
path: root/g4f
diff options
context:
space:
mode:
authorkqlio67 <kqlio67@users.noreply.github.com>2024-09-12 20:35:31 +0200
committerkqlio67 <kqlio67@users.noreply.github.com>2024-09-12 20:35:31 +0200
commitabea4ddbcca40c1c1da51507363867fb3664228f (patch)
tree20d907d877d826aa02f44ecee5005197f05cdd14 /g4f
parentAdded a new provider for generating images: g4f/Provider/Prodia.py (diff)
downloadgpt4free-abea4ddbcca40c1c1da51507363867fb3664228f.tar
gpt4free-abea4ddbcca40c1c1da51507363867fb3664228f.tar.gz
gpt4free-abea4ddbcca40c1c1da51507363867fb3664228f.tar.bz2
gpt4free-abea4ddbcca40c1c1da51507363867fb3664228f.tar.lz
gpt4free-abea4ddbcca40c1c1da51507363867fb3664228f.tar.xz
gpt4free-abea4ddbcca40c1c1da51507363867fb3664228f.tar.zst
gpt4free-abea4ddbcca40c1c1da51507363867fb3664228f.zip
Diffstat (limited to 'g4f')
-rw-r--r--g4f/Provider/HuggingChat.py1
-rw-r--r--g4f/Provider/HuggingFace.py23
2 files changed, 6 insertions, 18 deletions
diff --git a/g4f/Provider/HuggingChat.py b/g4f/Provider/HuggingChat.py
index fad44957..7edb2f9f 100644
--- a/g4f/Provider/HuggingChat.py
+++ b/g4f/Provider/HuggingChat.py
@@ -12,6 +12,7 @@ class HuggingChat(AbstractProvider, ProviderModelMixin):
working = True
supports_stream = True
default_model = "meta-llama/Meta-Llama-3.1-70B-Instruct"
+
models = [
default_model,
'CohereForAI/c4ai-command-r-plus-08-2024',
diff --git a/g4f/Provider/HuggingFace.py b/g4f/Provider/HuggingFace.py
index 4fe02739..586e5f5f 100644
--- a/g4f/Provider/HuggingFace.py
+++ b/g4f/Provider/HuggingFace.py
@@ -9,29 +9,16 @@ from .helper import get_connector
from ..errors import RateLimitError, ModelNotFoundError
from ..requests.raise_for_status import raise_for_status
+from .HuggingChat import HuggingChat
+
class HuggingFace(AsyncGeneratorProvider, ProviderModelMixin):
url = "https://huggingface.co/chat"
working = True
needs_auth = True
supports_message_history = True
- default_model = "meta-llama/Meta-Llama-3.1-70B-Instruct"
- models = [
- default_model,
- 'CohereForAI/c4ai-command-r-plus-08-2024',
- 'mistralai/Mixtral-8x7B-Instruct-v0.1',
- 'NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO',
- 'mistralai/Mistral-7B-Instruct-v0.3',
- 'microsoft/Phi-3-mini-4k-instruct',
- ]
-
- model_aliases = {
- "llama-3.1-70b": "meta-llama/Meta-Llama-3.1-70B-Instruct",
- "command-r-plus": "CohereForAI/c4ai-command-r-plus-08-2024",
- "mixtral-8x7b": "mistralai/Mixtral-8x7B-Instruct-v0.1",
- "mixtral-8x7b": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
- "mistral-7b": "mistralai/Mistral-7B-Instruct-v0.3",
- "phi-3-mini-4k": "microsoft/Phi-3-mini-4k-instruct",
- }
+ default_model = HuggingChat.default_model
+ models = HuggingChat.models
+ model_aliases = HuggingChat.model_aliases
@classmethod
def get_model(cls, model: str) -> str: