From 81cf5d7c771c4ff1a9d49db754ccef780f8b1d99 Mon Sep 17 00:00:00 2001 From: kqlio67 <166700875+kqlio67@users.noreply.github.com> Date: Fri, 19 Apr 2024 12:10:56 +0000 Subject: A few small fixes in GUI and Providers (#1861) --- g4f/Provider/Llama.py | 8 +++---- g4f/Provider/PerplexityLabs.py | 7 ++---- g4f/gui/client/index.html | 2 +- g4f/models.py | 53 +++++++++++++++++++++++++++--------------- 4 files changed, 41 insertions(+), 29 deletions(-) (limited to 'g4f') diff --git a/g4f/Provider/Llama.py b/g4f/Provider/Llama.py index 4d19866e..8f3e9ea2 100644 --- a/g4f/Provider/Llama.py +++ b/g4f/Provider/Llama.py @@ -16,12 +16,12 @@ class Llama(AsyncGeneratorProvider, ProviderModelMixin): "meta/llama-2-7b-chat", "meta/llama-2-13b-chat", "meta/llama-2-70b-chat", - "meta/llama-3-8b-chat", - "meta/llama-3-70b-chat", + "meta/meta-llama-3-8b-instruct", + "meta/meta-llama-3-70b-instruct", ] model_aliases = { - "meta-llama/Meta-Llama-3-8b": "meta/llama-3-8b-chat", - "meta-llama/Meta-Llama-3-70b": "meta/llama-3-70b-chat", + "meta-llama/Meta-Llama-3-8b-instruct": "meta/meta-llama-3-8b-instruct", + "meta-llama/Meta-Llama-3-70b-instruct": "meta/meta-llama-3-70b-instruct", "meta-llama/Llama-2-7b-chat-hf": "meta/llama-2-7b-chat", "meta-llama/Llama-2-13b-chat-hf": "meta/llama-2-13b-chat", "meta-llama/Llama-2-70b-chat-hf": "meta/llama-2-70b-chat", diff --git a/g4f/Provider/PerplexityLabs.py b/g4f/Provider/PerplexityLabs.py index ab36d284..b6fec53c 100644 --- a/g4f/Provider/PerplexityLabs.py +++ b/g4f/Provider/PerplexityLabs.py @@ -15,10 +15,7 @@ class PerplexityLabs(AsyncGeneratorProvider, ProviderModelMixin): working = True default_model = "mixtral-8x7b-instruct" models = [ - "sonar-small-online", "sonar-medium-online", "sonar-small-chat", "sonar-medium-chat", "mistral-7b-instruct", - "codellama-70b-instruct", "llava-v1.5-7b-wrapper", "llava-v1.6-34b", "mixtral-8x7b-instruct", - "gemma-2b-it", "gemma-7b-it" - "mistral-medium", "related", "dbrx-instruct" + "sonar-small-online", "sonar-medium-online", "sonar-small-chat", "sonar-medium-chat", "dbrx-instruct", "claude-3-haiku-20240307", "llama-3-8b-instruct", "llama-3-70b-instruct", "codellama-70b-instruct", "mistral-7b-instruct", "llava-v1.5-7b-wrapper", "llava-v1.6-34b", "mixtral-8x7b-instruct", "mixtral-8x22b-instruct", "mistral-medium", "gemma-2b-it", "gemma-7b-it", "related" ] model_aliases = { "mistralai/Mistral-7B-Instruct-v0.1": "mistral-7b-instruct", @@ -93,4 +90,4 @@ class PerplexityLabs(AsyncGeneratorProvider, ProviderModelMixin): if data["final"]: break except: - raise RuntimeError(f"Message: {message}") \ No newline at end of file + raise RuntimeError(f"Message: {message}") diff --git a/g4f/gui/client/index.html b/g4f/gui/client/index.html index 8668c21e..d84bbbe9 100644 --- a/g4f/gui/client/index.html +++ b/g4f/gui/client/index.html @@ -220,7 +220,7 @@ - + diff --git a/g4f/models.py b/g4f/models.py index 4af1c31e..2b7c69f3 100644 --- a/g4f/models.py +++ b/g4f/models.py @@ -4,28 +4,33 @@ from dataclasses import dataclass from .Provider import RetryProvider, ProviderType from .Provider import ( + Aichatos, + Bing, + Blackbox, Chatgpt4Online, - PerplexityLabs, - GeminiProChat, - ChatgptNext, - HuggingChat, - HuggingFace, - OpenaiChat, ChatgptAi, + ChatgptNext, + Cohere, + Cnote, DeepInfra, + Feedough, + FreeGpt, + Gemini, + GeminiProChat, GigaChat, + HuggingChat, + HuggingFace, + Koala, Liaobots, - FreeGpt, Llama, + OpenaiChat, + PerplexityLabs, + Pi, Vercel, - Gemini, - Koala, - Cohere, - Bing, You, - Pi, ) + @dataclass(unsafe_hash=True) class Model: """ @@ -79,6 +84,9 @@ gpt_35_turbo = Model( ChatgptNext, Koala, OpenaiChat, + Aichatos, + Cnote, + Feedough, ]) ) @@ -132,14 +140,14 @@ llama2_70b = Model( best_provider = RetryProvider([Llama, DeepInfra, HuggingChat]) ) -llama3_8b = Model( - name = "meta-llama/Meta-Llama-3-8b", +llama3_8b_instruct = Model( + name = "meta-llama/Meta-Llama-3-8b-instruct", base_provider = "meta", best_provider = RetryProvider([Llama]) ) -llama3_70b = Model( - name = "meta-llama/Meta-Llama-3-70b", +llama3_70b_instruct = Model( + name = "meta-llama/Meta-Llama-3-70b-instruct", base_provider = "meta", best_provider = RetryProvider([Llama, HuggingChat]) ) @@ -291,6 +299,12 @@ command_r_plus = Model( best_provider = RetryProvider([HuggingChat, Cohere]) ) +blackbox = Model( + name = 'blackbox', + base_provider = 'blackbox', + best_provider = Blackbox +) + class ModelUtils: """ Utility class for mapping string identifiers to Model instances. @@ -314,12 +328,12 @@ class ModelUtils: 'gpt-4-32k-0613' : gpt_4_32k_0613, 'gpt-4-turbo' : gpt_4_turbo, - # Llama 2 + # Llama 'llama2-7b' : llama2_7b, 'llama2-13b': llama2_13b, 'llama2-70b': llama2_70b, - 'llama3-8b' : llama3_8b, - 'llama3-70b': llama3_70b, + 'llama3-8b-instruct' : llama3_8b_instruct, + 'llama3-70b-instruct': llama3_70b_instruct, 'codellama-34b-instruct': codellama_34b_instruct, 'codellama-70b-instruct': codellama_70b_instruct, @@ -345,6 +359,7 @@ class ModelUtils: 'claude-3-sonnet': claude_3_sonnet, # other + 'blackbox': blackbox, 'command-r+': command_r_plus, 'dbrx-instruct': dbrx_instruct, 'lzlv-70b': lzlv_70b, -- cgit v1.2.3