summaryrefslogtreecommitdiffstats
path: root/g4f/models.py
diff options
context:
space:
mode:
authorkqlio67 <kqlio67@users.noreply.github.com>2024-11-09 22:44:52 +0100
committerkqlio67 <kqlio67@users.noreply.github.com>2024-11-09 22:44:52 +0100
commit3a15957d221a5e532ffe888e79a1fc2245c0fde1 (patch)
tree735297d8f8e5e5713aed4baf53c8ec694e3cb969 /g4f/models.py
parentUpdate (g4f/gui/client/index.html) (diff)
downloadgpt4free-3a15957d221a5e532ffe888e79a1fc2245c0fde1.tar
gpt4free-3a15957d221a5e532ffe888e79a1fc2245c0fde1.tar.gz
gpt4free-3a15957d221a5e532ffe888e79a1fc2245c0fde1.tar.bz2
gpt4free-3a15957d221a5e532ffe888e79a1fc2245c0fde1.tar.lz
gpt4free-3a15957d221a5e532ffe888e79a1fc2245c0fde1.tar.xz
gpt4free-3a15957d221a5e532ffe888e79a1fc2245c0fde1.tar.zst
gpt4free-3a15957d221a5e532ffe888e79a1fc2245c0fde1.zip
Diffstat (limited to 'g4f/models.py')
-rw-r--r--g4f/models.py244
1 files changed, 42 insertions, 202 deletions
diff --git a/g4f/models.py b/g4f/models.py
index 8788ab77..ec0ebd32 100644
--- a/g4f/models.py
+++ b/g4f/models.py
@@ -98,32 +98,32 @@ default = Model(
gpt_35_turbo = Model(
name = 'gpt-3.5-turbo',
base_provider = 'OpenAI',
- best_provider = IterListProvider([DarkAI, Airforce, Liaobots, Allyfy])
+ best_provider = IterListProvider([DarkAI, Liaobots, Allyfy])
)
# gpt-4
gpt_4o = Model(
name = 'gpt-4o',
base_provider = 'OpenAI',
- best_provider = IterListProvider([Blackbox, ChatGptEs, DarkAI, Airforce, ChatGpt, Liaobots, OpenaiChat])
+ best_provider = IterListProvider([Blackbox, ChatGptEs, DarkAI, ChatGpt, Airforce, Liaobots, OpenaiChat])
)
gpt_4o_mini = Model(
name = 'gpt-4o-mini',
base_provider = 'OpenAI',
- best_provider = IterListProvider([DDG, ChatGptEs, FreeNetfly, Pizzagpt, MagickPen, RubiksAI, Liaobots, ChatGpt, Airforce, OpenaiChat])
+ best_provider = IterListProvider([DDG, ChatGptEs, FreeNetfly, Pizzagpt, MagickPen, ChatGpt, Airforce, RubiksAI, Liaobots, OpenaiChat])
)
gpt_4_turbo = Model(
name = 'gpt-4-turbo',
base_provider = 'OpenAI',
- best_provider = IterListProvider([Liaobots, Airforce, ChatGpt, Bing])
+ best_provider = IterListProvider([ChatGpt, Airforce, Liaobots, Bing])
)
gpt_4 = Model(
name = 'gpt-4',
base_provider = 'OpenAI',
- best_provider = IterListProvider([Chatgpt4Online, Ai4Chat, ChatGpt, Airforce, Bing, OpenaiChat, gpt_4_turbo.best_provider, gpt_4o.best_provider, gpt_4o_mini.best_provider])
+ best_provider = IterListProvider([Chatgpt4Online, ChatGpt, Bing, OpenaiChat, gpt_4_turbo.best_provider, gpt_4o.best_provider, gpt_4o_mini.best_provider])
)
# o1
@@ -161,24 +161,17 @@ llama_2_7b = Model(
base_provider = "Meta Llama",
best_provider = IterListProvider([Cloudflare, Airforce])
)
-
-llama_2_13b = Model(
- name = "llama-2-13b",
- base_provider = "Meta Llama",
- best_provider = Airforce
-)
-
# llama 3
llama_3_8b = Model(
name = "llama-3-8b",
base_provider = "Meta Llama",
- best_provider = IterListProvider([Cloudflare, Airforce])
+ best_provider = IterListProvider([Cloudflare])
)
llama_3_70b = Model(
name = "llama-3-70b",
base_provider = "Meta Llama",
- best_provider = IterListProvider([ReplicateHome, Airforce])
+ best_provider = IterListProvider([ReplicateHome])
)
# llama 3.1
@@ -191,84 +184,39 @@ llama_3_1_8b = Model(
llama_3_1_70b = Model(
name = "llama-3.1-70b",
base_provider = "Meta Llama",
- best_provider = IterListProvider([DDG, DeepInfraChat, Blackbox, TeachAnything, DarkAI, AiMathGPT, RubiksAI, Airforce, HuggingChat, HuggingFace, PerplexityLabs])
+ best_provider = IterListProvider([DDG, DeepInfraChat, Blackbox, TeachAnything, DarkAI, AiMathGPT, Airforce, RubiksAI, HuggingChat, HuggingFace, PerplexityLabs])
)
llama_3_1_405b = Model(
name = "llama-3.1-405b",
base_provider = "Meta Llama",
- best_provider = IterListProvider([Blackbox, DarkAI, Airforce])
+ best_provider = IterListProvider([Blackbox, DarkAI])
)
# llama 3.2
llama_3_2_1b = Model(
name = "llama-3.2-1b",
base_provider = "Meta Llama",
- best_provider = IterListProvider([Cloudflare, Airforce])
-)
-
-llama_3_2_3b = Model(
- name = "llama-3.2-3b",
- base_provider = "Meta Llama",
- best_provider = IterListProvider([Airforce])
+ best_provider = IterListProvider([Cloudflare])
)
llama_3_2_11b = Model(
name = "llama-3.2-11b",
base_provider = "Meta Llama",
- best_provider = IterListProvider([HuggingChat, Airforce, HuggingFace])
-)
-
-llama_3_2_90b = Model(
- name = "llama-3.2-90b",
- base_provider = "Meta Llama",
- best_provider = IterListProvider([Airforce])
-)
-
-
-# llamaguard
-llamaguard_7b = Model(
- name = "llamaguard-7b",
- base_provider = "Meta Llama",
- best_provider = Airforce
-)
-
-llamaguard_2_8b = Model(
- name = "llamaguard-2-8b",
- base_provider = "Meta Llama",
- best_provider = Airforce
-)
-
-llamaguard_3_8b = Model(
- name = "llamaguard-3-8b",
- base_provider = "Meta Llama",
- best_provider = Airforce
-)
-
-llamaguard_3_11b = Model(
- name = "llamaguard-3-11b",
- base_provider = "Meta Llama",
- best_provider = Airforce
+ best_provider = IterListProvider([HuggingChat, HuggingFace])
)
-
### Mistral ###
mistral_7b = Model(
name = "mistral-7b",
base_provider = "Mistral",
- best_provider = IterListProvider([Free2GPT, Airforce])
+ best_provider = IterListProvider([Free2GPT])
)
mixtral_8x7b = Model(
name = "mixtral-8x7b",
base_provider = "Mistral",
- best_provider = IterListProvider([DDG, ReplicateHome, Airforce])
-)
-
-mixtral_8x22b = Model(
- name = "mixtral-8x22b",
- base_provider = "Mistral",
- best_provider = IterListProvider([Airforce])
+ best_provider = IterListProvider([DDG, ReplicateHome])
)
mistral_nemo = Model(
@@ -279,8 +227,8 @@ mistral_nemo = Model(
### NousResearch ###
-hermes_2 = Model(
- name = "hermes-2",
+hermes_2_pro = Model(
+ name = "hermes-2-pro",
base_provider = "NousResearch",
best_provider = Airforce
)
@@ -305,12 +253,6 @@ phi_2 = Model(
best_provider = IterListProvider([Cloudflare, Airforce])
)
-phi_3_medium_4k = Model(
- name = "phi-3-medium-4k",
- base_provider = "Microsoft",
- best_provider = None
-)
-
phi_3_5_mini = Model(
name = "phi-3.5-mini",
base_provider = "Microsoft",
@@ -322,13 +264,13 @@ phi_3_5_mini = Model(
gemini_pro = Model(
name = 'gemini-pro',
base_provider = 'Google DeepMind',
- best_provider = IterListProvider([GeminiPro, Blackbox, AIChatFree, FreeGpt, Airforce, Liaobots])
+ best_provider = IterListProvider([GeminiPro, Blackbox, AIChatFree, FreeGpt, Liaobots])
)
gemini_flash = Model(
name = 'gemini-flash',
base_provider = 'Google DeepMind',
- best_provider = IterListProvider([Blackbox, GizAI, Airforce, Liaobots])
+ best_provider = IterListProvider([Blackbox, GizAI, Liaobots])
)
gemini = Model(
@@ -341,13 +283,7 @@ gemini = Model(
gemma_2b = Model(
name = 'gemma-2b',
base_provider = 'Google',
- best_provider = IterListProvider([ReplicateHome, Airforce])
-)
-
-gemma_2b_27b = Model(
- name = 'gemma-2b-27b',
- base_provider = 'Google',
- best_provider = IterListProvider([Airforce])
+ best_provider = IterListProvider([ReplicateHome])
)
gemma_7b = Model(
@@ -356,13 +292,6 @@ gemma_7b = Model(
best_provider = Cloudflare
)
-# gemma 2
-gemma_2_9b = Model(
- name = 'gemma-2-9b',
- base_provider = 'Google',
- best_provider = Airforce
-)
-
### Anthropic ###
claude_2_1 = Model(
@@ -419,15 +348,6 @@ blackboxai_pro = Model(
best_provider = Blackbox
)
-
-### Databricks ###
-dbrx_instruct = Model(
- name = 'dbrx-instruct',
- base_provider = 'Databricks',
- best_provider = IterListProvider([Airforce])
-)
-
-
### CohereForAI ###
command_r_plus = Model(
name = 'command-r-plus',
@@ -466,28 +386,10 @@ qwen_1_5_14b = Model(
qwen_2_72b = Model(
name = 'qwen-2-72b',
base_provider = 'Qwen',
- best_provider = IterListProvider([DeepInfraChat, HuggingChat, Airforce, HuggingFace])
-)
-
-qwen_2_5_7b = Model(
- name = 'qwen-2-5-7b',
- base_provider = 'Qwen',
- best_provider = Airforce
-)
-
-qwen_2_5_72b = Model(
- name = 'qwen-2-5-72b',
- base_provider = 'Qwen',
- best_provider = Airforce
+ best_provider = IterListProvider([DeepInfraChat, HuggingChat, HuggingFace])
)
### Upstage ###
-solar_10_7b = Model(
- name = 'solar-10-7b',
- base_provider = 'Upstage',
- best_provider = Airforce
-)
-
solar_mini = Model(
name = 'solar-mini',
base_provider = 'Upstage',
@@ -519,7 +421,7 @@ deepseek_coder = Model(
wizardlm_2_8x22b = Model(
name = 'wizardlm-2-8x22b',
base_provider = 'WizardLM',
- best_provider = IterListProvider([DeepInfraChat, Airforce])
+ best_provider = IterListProvider([DeepInfraChat])
)
### Yorickvp ###
@@ -529,44 +431,11 @@ llava_13b = Model(
best_provider = ReplicateHome
)
-
-### OpenBMB ###
-minicpm_llama_3_v2_5 = Model(
- name = 'minicpm-llama-3-v2.5',
- base_provider = 'OpenBMB',
- best_provider = None
-)
-
-
-### Lzlv ###
-lzlv_70b = Model(
- name = 'lzlv-70b',
- base_provider = 'Lzlv',
- best_provider = None
-)
-
-
### OpenChat ###
-openchat_3_6_8b = Model(
- name = 'openchat-3.6-8b',
+openchat_3_5 = Model(
+ name = 'openchat-3.5',
base_provider = 'OpenChat',
- best_provider = None
-)
-
-
-### Phind ###
-phind_codellama_34b_v2 = Model(
- name = 'phind-codellama-34b-v2',
- base_provider = 'Phind',
- best_provider = None
-)
-
-
-### Cognitive Computations ###
-dolphin_2_9_1_llama_3_70b = Model(
- name = 'dolphin-2.9.1-llama-3-70b',
- base_provider = 'Cognitive Computations',
- best_provider = None
+ best_provider = Airforce
)
@@ -650,6 +519,13 @@ zephyr_7b = Model(
best_provider = Airforce
)
+### Inferless ###
+neural_7b = Model(
+ name = 'neural-7b',
+ base_provider = 'inferless',
+ best_provider = Airforce
+)
+
#############
@@ -660,7 +536,7 @@ zephyr_7b = Model(
sdxl = Model(
name = 'sdxl',
base_provider = 'Stability AI',
- best_provider = IterListProvider([ReplicateHome, Airforce])
+ best_provider = IterListProvider([ReplicateHome])
)
@@ -740,7 +616,7 @@ flux_4o = Model(
flux_schnell = Model(
name = 'flux-schnell',
base_provider = 'Flux AI',
- best_provider = IterListProvider([ReplicateHome])
+ best_provider = ReplicateHome
)
@@ -786,7 +662,6 @@ class ModelUtils:
# llama-2
'llama-2-7b': llama_2_7b,
-'llama-2-13b': llama_2_13b,
# llama-3
'llama-3-8b': llama_3_8b,
@@ -799,33 +674,23 @@ class ModelUtils:
# llama-3.2
'llama-3.2-1b': llama_3_2_1b,
-'llama-3.2-3b': llama_3_2_3b,
'llama-3.2-11b': llama_3_2_11b,
-'llama-3.2-90b': llama_3_2_90b,
-
-# llamaguard
-'llamaguard-7b': llamaguard_7b,
-'llamaguard-2-8b': llamaguard_2_8b,
-'llamaguard-3-8b': llamaguard_3_8b,
-'llamaguard-3-11b': llamaguard_3_11b,
-
+
### Mistral ###
'mistral-7b': mistral_7b,
'mixtral-8x7b': mixtral_8x7b,
-'mixtral-8x22b': mixtral_8x22b,
'mistral-nemo': mistral_nemo,
### NousResearch ###
-'hermes-2': hermes_2,
+'hermes-2-pro': hermes_2_pro,
'hermes-2-dpo': hermes_2_dpo,
'hermes-3': hermes_3,
### Microsoft ###
'phi-2': phi_2,
-'phi_3_medium-4k': phi_3_medium_4k,
'phi-3.5-mini': phi_3_5_mini,
@@ -837,12 +702,8 @@ class ModelUtils:
# gemma
'gemma-2b': gemma_2b,
-'gemma-2b-27b': gemma_2b_27b,
'gemma-7b': gemma_7b,
-# gemma-2
-'gemma-2-9b': gemma_2_9b,
-
### Anthropic ###
'claude-2.1': claude_2_1,
@@ -868,10 +729,6 @@ class ModelUtils:
### CohereForAI ###
'command-r+': command_r_plus,
-
-### Databricks ###
-'dbrx-instruct': dbrx_instruct,
-
### GigaChat ###
'gigachat': gigachat,
@@ -887,14 +744,9 @@ class ModelUtils:
# qwen 2
'qwen-2-72b': qwen_2_72b,
-
-# qwen 2-5
-'qwen-2-5-7b': qwen_2_5_7b,
-'qwen-2-5-72b': qwen_2_5_72b,
-
+
### Upstage ###
-'solar-10-7b': solar_10_7b,
'solar-mini': solar_mini,
'solar-pro': solar_pro,
@@ -913,27 +765,11 @@ class ModelUtils:
### WizardLM ###
'wizardlm-2-8x22b': wizardlm_2_8x22b,
-
-
-### OpenBMB ###
-'minicpm-llama-3-v2.5': minicpm_llama_3_v2_5,
-
-
-### Lzlv ###
-'lzlv-70b': lzlv_70b,
-
+
### OpenChat ###
-'openchat-3.6-8b': openchat_3_6_8b,
-
-
-### Phind ###
-'phind-codellama-34b-v2': phind_codellama_34b_v2,
-
-
-### Cognitive Computations ###
-'dolphin-2.9.1-llama-3-70b': dolphin_2_9_1_llama_3_70b,
-
+'openchat-3.5': openchat_3_5,
+
### x.ai ###
'grok-2': grok_2,
@@ -972,6 +808,10 @@ class ModelUtils:
### HuggingFaceH4 ###
'zephyr-7b': zephyr_7b,
+
+
+### Inferless ###
+'neural-7b': neural_7b,