summaryrefslogtreecommitdiffstats
path: root/g4f
diff options
context:
space:
mode:
authorkqlio67 <kqlio67@users.noreply.github.com>2024-11-09 22:44:52 +0100
committerkqlio67 <kqlio67@users.noreply.github.com>2024-11-09 22:44:52 +0100
commit3a15957d221a5e532ffe888e79a1fc2245c0fde1 (patch)
tree735297d8f8e5e5713aed4baf53c8ec694e3cb969 /g4f
parentUpdate (g4f/gui/client/index.html) (diff)
downloadgpt4free-3a15957d221a5e532ffe888e79a1fc2245c0fde1.tar
gpt4free-3a15957d221a5e532ffe888e79a1fc2245c0fde1.tar.gz
gpt4free-3a15957d221a5e532ffe888e79a1fc2245c0fde1.tar.bz2
gpt4free-3a15957d221a5e532ffe888e79a1fc2245c0fde1.tar.lz
gpt4free-3a15957d221a5e532ffe888e79a1fc2245c0fde1.tar.xz
gpt4free-3a15957d221a5e532ffe888e79a1fc2245c0fde1.tar.zst
gpt4free-3a15957d221a5e532ffe888e79a1fc2245c0fde1.zip
Diffstat (limited to 'g4f')
-rw-r--r--g4f/Provider/airforce/AirforceChat.py261
-rw-r--r--g4f/models.py244
2 files changed, 69 insertions, 436 deletions
diff --git a/g4f/Provider/airforce/AirforceChat.py b/g4f/Provider/airforce/AirforceChat.py
index 8affbe5c..63a0460f 100644
--- a/g4f/Provider/airforce/AirforceChat.py
+++ b/g4f/Provider/airforce/AirforceChat.py
@@ -3,6 +3,7 @@ import re
from aiohttp import ClientSession
import json
from typing import List
+import requests
from ...typing import AsyncResult, Messages
from ..base_provider import AsyncGeneratorProvider, ProviderModelMixin
@@ -51,258 +52,50 @@ class AirforceChat(AsyncGeneratorProvider, ProviderModelMixin):
supports_system_message = True
supports_message_history = True
- default_model = 'llama-3-70b-chat'
- text_models = [
- # anthropic
- 'claude-3-haiku-20240307',
- 'claude-3-sonnet-20240229',
- 'claude-3-5-sonnet-20240620',
- 'claude-3-5-sonnet-20241022',
- 'claude-3-opus-20240229',
-
- # openai
- 'chatgpt-4o-latest',
- 'gpt-4',
- 'gpt-4-turbo',
- 'gpt-4o-2024-05-13',
- 'gpt-4o-mini-2024-07-18',
- 'gpt-4o-mini',
- 'gpt-4o-2024-08-06',
- 'gpt-3.5-turbo',
- 'gpt-3.5-turbo-0125',
- 'gpt-3.5-turbo-1106',
- 'gpt-4o',
- 'gpt-4-turbo-2024-04-09',
- 'gpt-4-0125-preview',
- 'gpt-4-1106-preview',
-
- # meta-llama
- default_model,
- 'llama-3-70b-chat-turbo',
- 'llama-3-8b-chat',
- 'llama-3-8b-chat-turbo',
- 'llama-3-70b-chat-lite',
- 'llama-3-8b-chat-lite',
- 'llama-2-13b-chat',
- 'llama-3.1-405b-turbo',
- 'llama-3.1-70b-turbo',
- 'llama-3.1-8b-turbo',
- 'LlamaGuard-2-8b',
- 'llamaguard-7b',
- 'Llama-Vision-Free',
- 'Llama-Guard-7b',
- 'Llama-3.2-90B-Vision-Instruct-Turbo',
- 'Meta-Llama-Guard-3-8B',
- 'Llama-3.2-11B-Vision-Instruct-Turbo',
- 'Llama-Guard-3-11B-Vision-Turbo',
- 'Llama-3.2-3B-Instruct-Turbo',
- 'Llama-3.2-1B-Instruct-Turbo',
- 'llama-2-7b-chat-int8',
- 'llama-2-7b-chat-fp16',
- 'Llama 3.1 405B Instruct',
- 'Llama 3.1 70B Instruct',
- 'Llama 3.1 8B Instruct',
-
- # mistral-ai
- 'Mixtral-8x7B-Instruct-v0.1',
- 'Mixtral-8x22B-Instruct-v0.1',
- 'Mistral-7B-Instruct-v0.1',
- 'Mistral-7B-Instruct-v0.2',
- 'Mistral-7B-Instruct-v0.3',
-
- # Gryphe
- 'MythoMax-L2-13b-Lite',
- 'MythoMax-L2-13b',
-
- # openchat
- 'openchat-3.5-0106',
-
- # qwen
- #'Qwen1.5-72B-Chat', # Empty answer
- #'Qwen1.5-110B-Chat', # Empty answer
- 'Qwen2-72B-Instruct',
- 'Qwen2.5-7B-Instruct-Turbo',
- 'Qwen2.5-72B-Instruct-Turbo',
-
- # google
- 'gemma-2b-it',
- 'gemma-2-9b-it',
- 'gemma-2-27b-it',
-
- # gemini
- 'gemini-1.5-flash',
- 'gemini-1.5-pro',
-
- # databricks
- 'dbrx-instruct',
-
- # deepseek-ai
- 'deepseek-coder-6.7b-base',
- 'deepseek-coder-6.7b-instruct',
- 'deepseek-math-7b-instruct',
-
- # NousResearch
- 'deepseek-math-7b-instruct',
- 'Nous-Hermes-2-Mixtral-8x7B-DPO',
- 'hermes-2-pro-mistral-7b',
-
- # teknium
- 'openhermes-2.5-mistral-7b',
-
- # microsoft
- 'WizardLM-2-8x22B',
- 'phi-2',
-
- # upstage
- 'SOLAR-10.7B-Instruct-v1.0',
-
- # pawan
- 'cosmosrp',
-
- # liquid
- 'lfm-40b-moe',
-
- # DiscoResearch
- 'discolm-german-7b-v1',
-
- # tiiuae
- 'falcon-7b-instruct',
-
- # defog
- 'sqlcoder-7b-2',
-
- # tinyllama
- 'tinyllama-1.1b-chat',
-
- # HuggingFaceH4
- 'zephyr-7b-beta',
- ]
+ default_model = 'llama-3.1-70b-chat'
+ response = requests.get('https://api.airforce/models')
+ data = response.json()
+
+ text_models = [model['id'] for model in data['data']]
models = [*text_models]
model_aliases = {
- # anthropic
- "claude-3-haiku": "claude-3-haiku-20240307",
- "claude-3-sonnet": "claude-3-sonnet-20240229",
- "claude-3.5-sonnet": "claude-3-5-sonnet-20240620",
- "claude-3.5-sonnet": "claude-3-5-sonnet-20241022",
- "claude-3-opus": "claude-3-opus-20240229",
-
- # openai
- "gpt-4o": "chatgpt-4o-latest",
- #"gpt-4": "gpt-4",
- #"gpt-4-turbo": "gpt-4-turbo",
- "gpt-4o": "gpt-4o-2024-05-13",
- "gpt-4o-mini": "gpt-4o-mini-2024-07-18",
- #"gpt-4o-mini": "gpt-4o-mini",
- "gpt-4o": "gpt-4o-2024-08-06",
- "gpt-3.5-turbo": "gpt-3.5-turbo",
- "gpt-3.5-turbo": "gpt-3.5-turbo-0125",
- "gpt-3.5-turbo": "gpt-3.5-turbo-1106",
- #"gpt-4o": "gpt-4o",
- "gpt-4-turbo": "gpt-4-turbo-2024-04-09",
- "gpt-4": "gpt-4-0125-preview",
- "gpt-4": "gpt-4-1106-preview",
-
- # meta-llama
- "llama-3-70b": "llama-3-70b-chat",
- "llama-3-8b": "llama-3-8b-chat",
- "llama-3-8b": "llama-3-8b-chat-turbo",
- "llama-3-70b": "llama-3-70b-chat-lite",
- "llama-3-8b": "llama-3-8b-chat-lite",
- "llama-2-13b": "llama-2-13b-chat",
- "llama-3.1-405b": "llama-3.1-405b-turbo",
- "llama-3.1-70b": "llama-3.1-70b-turbo",
- "llama-3.1-8b": "llama-3.1-8b-turbo",
- "llamaguard-2-8b": "LlamaGuard-2-8b",
- "llamaguard-7b": "llamaguard-7b",
- #"llama_vision_free": "Llama-Vision-Free", # Unknown
- "llamaguard-7b": "Llama-Guard-7b",
- "llama-3.2-90b": "Llama-3.2-90B-Vision-Instruct-Turbo",
- "llamaguard-3-8b": "Meta-Llama-Guard-3-8B",
- "llama-3.2-11b": "Llama-3.2-11B-Vision-Instruct-Turbo",
- "llamaguard-3-11b": "Llama-Guard-3-11B-Vision-Turbo",
- "llama-3.2-3b": "Llama-3.2-3B-Instruct-Turbo",
- "llama-3.2-1b": "Llama-3.2-1B-Instruct-Turbo",
- "llama-2-7b": "llama-2-7b-chat-int8",
- "llama-2-7b": "llama-2-7b-chat-fp16",
- "llama-3.1-405b": "Llama 3.1 405B Instruct",
- "llama-3.1-70b": "Llama 3.1 70B Instruct",
- "llama-3.1-8b": "Llama 3.1 8B Instruct",
-
- # mistral-ai
- "mixtral-8x7b": "Mixtral-8x7B-Instruct-v0.1",
- "mixtral-8x22b": "Mixtral-8x22B-Instruct-v0.1",
- "mixtral-8x7b": "Mistral-7B-Instruct-v0.1",
- "mixtral-8x7b": "Mistral-7B-Instruct-v0.2",
- "mixtral-8x7b": "Mistral-7B-Instruct-v0.3",
-
- # Gryphe
- "mythomax-13b": "MythoMax-L2-13b-Lite",
- "mythomax-13b": "MythoMax-L2-13b",
-
# openchat
"openchat-3.5": "openchat-3.5-0106",
-
- # qwen
- #"qwen-1.5-72b": "Qwen1.5-72B-Chat", # Empty answer
- #"qwen-1.5-110b": "Qwen1.5-110B-Chat", # Empty answer
- "qwen-2-72b": "Qwen2-72B-Instruct",
- "qwen-2-5-7b": "Qwen2.5-7B-Instruct-Turbo",
- "qwen-2-5-72b": "Qwen2.5-72B-Instruct-Turbo",
-
- # google
- "gemma-2b": "gemma-2b-it",
- "gemma-2-9b": "gemma-2-9b-it",
- "gemma-2b-27b": "gemma-2-27b-it",
-
- # gemini
- "gemini-flash": "gemini-1.5-flash",
- "gemini-pro": "gemini-1.5-pro",
-
- # databricks
- "dbrx-instruct": "dbrx-instruct",
-
+
# deepseek-ai
- #"deepseek-coder": "deepseek-coder-6.7b-base",
"deepseek-coder": "deepseek-coder-6.7b-instruct",
- #"deepseek-math": "deepseek-math-7b-instruct",
-
+
# NousResearch
- #"deepseek-math": "deepseek-math-7b-instruct",
"hermes-2-dpo": "Nous-Hermes-2-Mixtral-8x7B-DPO",
- "hermes-2": "hermes-2-pro-mistral-7b",
-
+ "hermes-2-pro": "hermes-2-pro-mistral-7b",
+
# teknium
"openhermes-2.5": "openhermes-2.5-mistral-7b",
-
- # microsoft
- "wizardlm-2-8x22b": "WizardLM-2-8x22B",
- #"phi-2": "phi-2",
-
- # upstage
- "solar-10-7b": "SOLAR-10.7B-Instruct-v1.0",
-
- # pawan
- #"cosmosrp": "cosmosrp",
-
+
# liquid
"lfm-40b": "lfm-40b-moe",
-
+
# DiscoResearch
"german-7b": "discolm-german-7b-v1",
-
- # tiiuae
- #"falcon-7b": "falcon-7b-instruct",
-
- # defog
- #"sqlcoder-7b": "sqlcoder-7b-2",
-
- # tinyllama
- #"tinyllama-1b": "tinyllama-1.1b-chat",
-
+
+ # meta-llama
+ "llama-2-7b": "llama-2-7b-chat-int8",
+ "llama-2-7b": "llama-2-7b-chat-fp16",
+ "llama-3.1-70b": "llama-3.1-70b-chat",
+ "llama-3.1-8b": "llama-3.1-8b-chat",
+ "llama-3.1-70b": "llama-3.1-70b-turbo",
+ "llama-3.1-8b": "llama-3.1-8b-turbo",
+
+ # inferless
+ "neural-7b": "neural-chat-7b-v3-1",
+
# HuggingFaceH4
"zephyr-7b": "zephyr-7b-beta",
+
+ # llmplayground.net
+ #"any-uncensored": "any-uncensored",
}
@classmethod
diff --git a/g4f/models.py b/g4f/models.py
index 8788ab77..ec0ebd32 100644
--- a/g4f/models.py
+++ b/g4f/models.py
@@ -98,32 +98,32 @@ default = Model(
gpt_35_turbo = Model(
name = 'gpt-3.5-turbo',
base_provider = 'OpenAI',
- best_provider = IterListProvider([DarkAI, Airforce, Liaobots, Allyfy])
+ best_provider = IterListProvider([DarkAI, Liaobots, Allyfy])
)
# gpt-4
gpt_4o = Model(
name = 'gpt-4o',
base_provider = 'OpenAI',
- best_provider = IterListProvider([Blackbox, ChatGptEs, DarkAI, Airforce, ChatGpt, Liaobots, OpenaiChat])
+ best_provider = IterListProvider([Blackbox, ChatGptEs, DarkAI, ChatGpt, Airforce, Liaobots, OpenaiChat])
)
gpt_4o_mini = Model(
name = 'gpt-4o-mini',
base_provider = 'OpenAI',
- best_provider = IterListProvider([DDG, ChatGptEs, FreeNetfly, Pizzagpt, MagickPen, RubiksAI, Liaobots, ChatGpt, Airforce, OpenaiChat])
+ best_provider = IterListProvider([DDG, ChatGptEs, FreeNetfly, Pizzagpt, MagickPen, ChatGpt, Airforce, RubiksAI, Liaobots, OpenaiChat])
)
gpt_4_turbo = Model(
name = 'gpt-4-turbo',
base_provider = 'OpenAI',
- best_provider = IterListProvider([Liaobots, Airforce, ChatGpt, Bing])
+ best_provider = IterListProvider([ChatGpt, Airforce, Liaobots, Bing])
)
gpt_4 = Model(
name = 'gpt-4',
base_provider = 'OpenAI',
- best_provider = IterListProvider([Chatgpt4Online, Ai4Chat, ChatGpt, Airforce, Bing, OpenaiChat, gpt_4_turbo.best_provider, gpt_4o.best_provider, gpt_4o_mini.best_provider])
+ best_provider = IterListProvider([Chatgpt4Online, ChatGpt, Bing, OpenaiChat, gpt_4_turbo.best_provider, gpt_4o.best_provider, gpt_4o_mini.best_provider])
)
# o1
@@ -161,24 +161,17 @@ llama_2_7b = Model(
base_provider = "Meta Llama",
best_provider = IterListProvider([Cloudflare, Airforce])
)
-
-llama_2_13b = Model(
- name = "llama-2-13b",
- base_provider = "Meta Llama",
- best_provider = Airforce
-)
-
# llama 3
llama_3_8b = Model(
name = "llama-3-8b",
base_provider = "Meta Llama",
- best_provider = IterListProvider([Cloudflare, Airforce])
+ best_provider = IterListProvider([Cloudflare])
)
llama_3_70b = Model(
name = "llama-3-70b",
base_provider = "Meta Llama",
- best_provider = IterListProvider([ReplicateHome, Airforce])
+ best_provider = IterListProvider([ReplicateHome])
)
# llama 3.1
@@ -191,84 +184,39 @@ llama_3_1_8b = Model(
llama_3_1_70b = Model(
name = "llama-3.1-70b",
base_provider = "Meta Llama",
- best_provider = IterListProvider([DDG, DeepInfraChat, Blackbox, TeachAnything, DarkAI, AiMathGPT, RubiksAI, Airforce, HuggingChat, HuggingFace, PerplexityLabs])
+ best_provider = IterListProvider([DDG, DeepInfraChat, Blackbox, TeachAnything, DarkAI, AiMathGPT, Airforce, RubiksAI, HuggingChat, HuggingFace, PerplexityLabs])
)
llama_3_1_405b = Model(
name = "llama-3.1-405b",
base_provider = "Meta Llama",
- best_provider = IterListProvider([Blackbox, DarkAI, Airforce])
+ best_provider = IterListProvider([Blackbox, DarkAI])
)
# llama 3.2
llama_3_2_1b = Model(
name = "llama-3.2-1b",
base_provider = "Meta Llama",
- best_provider = IterListProvider([Cloudflare, Airforce])
-)
-
-llama_3_2_3b = Model(
- name = "llama-3.2-3b",
- base_provider = "Meta Llama",
- best_provider = IterListProvider([Airforce])
+ best_provider = IterListProvider([Cloudflare])
)
llama_3_2_11b = Model(
name = "llama-3.2-11b",
base_provider = "Meta Llama",
- best_provider = IterListProvider([HuggingChat, Airforce, HuggingFace])
-)
-
-llama_3_2_90b = Model(
- name = "llama-3.2-90b",
- base_provider = "Meta Llama",
- best_provider = IterListProvider([Airforce])
-)
-
-
-# llamaguard
-llamaguard_7b = Model(
- name = "llamaguard-7b",
- base_provider = "Meta Llama",
- best_provider = Airforce
-)
-
-llamaguard_2_8b = Model(
- name = "llamaguard-2-8b",
- base_provider = "Meta Llama",
- best_provider = Airforce
-)
-
-llamaguard_3_8b = Model(
- name = "llamaguard-3-8b",
- base_provider = "Meta Llama",
- best_provider = Airforce
-)
-
-llamaguard_3_11b = Model(
- name = "llamaguard-3-11b",
- base_provider = "Meta Llama",
- best_provider = Airforce
+ best_provider = IterListProvider([HuggingChat, HuggingFace])
)
-
### Mistral ###
mistral_7b = Model(
name = "mistral-7b",
base_provider = "Mistral",
- best_provider = IterListProvider([Free2GPT, Airforce])
+ best_provider = IterListProvider([Free2GPT])
)
mixtral_8x7b = Model(
name = "mixtral-8x7b",
base_provider = "Mistral",
- best_provider = IterListProvider([DDG, ReplicateHome, Airforce])
-)
-
-mixtral_8x22b = Model(
- name = "mixtral-8x22b",
- base_provider = "Mistral",
- best_provider = IterListProvider([Airforce])
+ best_provider = IterListProvider([DDG, ReplicateHome])
)
mistral_nemo = Model(
@@ -279,8 +227,8 @@ mistral_nemo = Model(
### NousResearch ###
-hermes_2 = Model(
- name = "hermes-2",
+hermes_2_pro = Model(
+ name = "hermes-2-pro",
base_provider = "NousResearch",
best_provider = Airforce
)
@@ -305,12 +253,6 @@ phi_2 = Model(
best_provider = IterListProvider([Cloudflare, Airforce])
)
-phi_3_medium_4k = Model(
- name = "phi-3-medium-4k",
- base_provider = "Microsoft",
- best_provider = None
-)
-
phi_3_5_mini = Model(
name = "phi-3.5-mini",
base_provider = "Microsoft",
@@ -322,13 +264,13 @@ phi_3_5_mini = Model(
gemini_pro = Model(
name = 'gemini-pro',
base_provider = 'Google DeepMind',
- best_provider = IterListProvider([GeminiPro, Blackbox, AIChatFree, FreeGpt, Airforce, Liaobots])
+ best_provider = IterListProvider([GeminiPro, Blackbox, AIChatFree, FreeGpt, Liaobots])
)
gemini_flash = Model(
name = 'gemini-flash',
base_provider = 'Google DeepMind',
- best_provider = IterListProvider([Blackbox, GizAI, Airforce, Liaobots])
+ best_provider = IterListProvider([Blackbox, GizAI, Liaobots])
)
gemini = Model(
@@ -341,13 +283,7 @@ gemini = Model(
gemma_2b = Model(
name = 'gemma-2b',
base_provider = 'Google',
- best_provider = IterListProvider([ReplicateHome, Airforce])
-)
-
-gemma_2b_27b = Model(
- name = 'gemma-2b-27b',
- base_provider = 'Google',
- best_provider = IterListProvider([Airforce])
+ best_provider = IterListProvider([ReplicateHome])
)
gemma_7b = Model(
@@ -356,13 +292,6 @@ gemma_7b = Model(
best_provider = Cloudflare
)
-# gemma 2
-gemma_2_9b = Model(
- name = 'gemma-2-9b',
- base_provider = 'Google',
- best_provider = Airforce
-)
-
### Anthropic ###
claude_2_1 = Model(
@@ -419,15 +348,6 @@ blackboxai_pro = Model(
best_provider = Blackbox
)
-
-### Databricks ###
-dbrx_instruct = Model(
- name = 'dbrx-instruct',
- base_provider = 'Databricks',
- best_provider = IterListProvider([Airforce])
-)
-
-
### CohereForAI ###
command_r_plus = Model(
name = 'command-r-plus',
@@ -466,28 +386,10 @@ qwen_1_5_14b = Model(
qwen_2_72b = Model(
name = 'qwen-2-72b',
base_provider = 'Qwen',
- best_provider = IterListProvider([DeepInfraChat, HuggingChat, Airforce, HuggingFace])
-)
-
-qwen_2_5_7b = Model(
- name = 'qwen-2-5-7b',
- base_provider = 'Qwen',
- best_provider = Airforce
-)
-
-qwen_2_5_72b = Model(
- name = 'qwen-2-5-72b',
- base_provider = 'Qwen',
- best_provider = Airforce
+ best_provider = IterListProvider([DeepInfraChat, HuggingChat, HuggingFace])
)
### Upstage ###
-solar_10_7b = Model(
- name = 'solar-10-7b',
- base_provider = 'Upstage',
- best_provider = Airforce
-)
-
solar_mini = Model(
name = 'solar-mini',
base_provider = 'Upstage',
@@ -519,7 +421,7 @@ deepseek_coder = Model(
wizardlm_2_8x22b = Model(
name = 'wizardlm-2-8x22b',
base_provider = 'WizardLM',
- best_provider = IterListProvider([DeepInfraChat, Airforce])
+ best_provider = IterListProvider([DeepInfraChat])
)
### Yorickvp ###
@@ -529,44 +431,11 @@ llava_13b = Model(
best_provider = ReplicateHome
)
-
-### OpenBMB ###
-minicpm_llama_3_v2_5 = Model(
- name = 'minicpm-llama-3-v2.5',
- base_provider = 'OpenBMB',
- best_provider = None
-)
-
-
-### Lzlv ###
-lzlv_70b = Model(
- name = 'lzlv-70b',
- base_provider = 'Lzlv',
- best_provider = None
-)
-
-
### OpenChat ###
-openchat_3_6_8b = Model(
- name = 'openchat-3.6-8b',
+openchat_3_5 = Model(
+ name = 'openchat-3.5',
base_provider = 'OpenChat',
- best_provider = None
-)
-
-
-### Phind ###
-phind_codellama_34b_v2 = Model(
- name = 'phind-codellama-34b-v2',
- base_provider = 'Phind',
- best_provider = None
-)
-
-
-### Cognitive Computations ###
-dolphin_2_9_1_llama_3_70b = Model(
- name = 'dolphin-2.9.1-llama-3-70b',
- base_provider = 'Cognitive Computations',
- best_provider = None
+ best_provider = Airforce
)
@@ -650,6 +519,13 @@ zephyr_7b = Model(
best_provider = Airforce
)
+### Inferless ###
+neural_7b = Model(
+ name = 'neural-7b',
+ base_provider = 'inferless',
+ best_provider = Airforce
+)
+
#############
@@ -660,7 +536,7 @@ zephyr_7b = Model(
sdxl = Model(
name = 'sdxl',
base_provider = 'Stability AI',
- best_provider = IterListProvider([ReplicateHome, Airforce])
+ best_provider = IterListProvider([ReplicateHome])
)
@@ -740,7 +616,7 @@ flux_4o = Model(
flux_schnell = Model(
name = 'flux-schnell',
base_provider = 'Flux AI',
- best_provider = IterListProvider([ReplicateHome])
+ best_provider = ReplicateHome
)
@@ -786,7 +662,6 @@ class ModelUtils:
# llama-2
'llama-2-7b': llama_2_7b,
-'llama-2-13b': llama_2_13b,
# llama-3
'llama-3-8b': llama_3_8b,
@@ -799,33 +674,23 @@ class ModelUtils:
# llama-3.2
'llama-3.2-1b': llama_3_2_1b,
-'llama-3.2-3b': llama_3_2_3b,
'llama-3.2-11b': llama_3_2_11b,
-'llama-3.2-90b': llama_3_2_90b,
-
-# llamaguard
-'llamaguard-7b': llamaguard_7b,
-'llamaguard-2-8b': llamaguard_2_8b,
-'llamaguard-3-8b': llamaguard_3_8b,
-'llamaguard-3-11b': llamaguard_3_11b,
-
+
### Mistral ###
'mistral-7b': mistral_7b,
'mixtral-8x7b': mixtral_8x7b,
-'mixtral-8x22b': mixtral_8x22b,
'mistral-nemo': mistral_nemo,
### NousResearch ###
-'hermes-2': hermes_2,
+'hermes-2-pro': hermes_2_pro,
'hermes-2-dpo': hermes_2_dpo,
'hermes-3': hermes_3,
### Microsoft ###
'phi-2': phi_2,
-'phi_3_medium-4k': phi_3_medium_4k,
'phi-3.5-mini': phi_3_5_mini,
@@ -837,12 +702,8 @@ class ModelUtils:
# gemma
'gemma-2b': gemma_2b,
-'gemma-2b-27b': gemma_2b_27b,
'gemma-7b': gemma_7b,
-# gemma-2
-'gemma-2-9b': gemma_2_9b,
-
### Anthropic ###
'claude-2.1': claude_2_1,
@@ -868,10 +729,6 @@ class ModelUtils:
### CohereForAI ###
'command-r+': command_r_plus,
-
-### Databricks ###
-'dbrx-instruct': dbrx_instruct,
-
### GigaChat ###
'gigachat': gigachat,
@@ -887,14 +744,9 @@ class ModelUtils:
# qwen 2
'qwen-2-72b': qwen_2_72b,
-
-# qwen 2-5
-'qwen-2-5-7b': qwen_2_5_7b,
-'qwen-2-5-72b': qwen_2_5_72b,
-
+
### Upstage ###
-'solar-10-7b': solar_10_7b,
'solar-mini': solar_mini,
'solar-pro': solar_pro,
@@ -913,27 +765,11 @@ class ModelUtils:
### WizardLM ###
'wizardlm-2-8x22b': wizardlm_2_8x22b,
-
-
-### OpenBMB ###
-'minicpm-llama-3-v2.5': minicpm_llama_3_v2_5,
-
-
-### Lzlv ###
-'lzlv-70b': lzlv_70b,
-
+
### OpenChat ###
-'openchat-3.6-8b': openchat_3_6_8b,
-
-
-### Phind ###
-'phind-codellama-34b-v2': phind_codellama_34b_v2,
-
-
-### Cognitive Computations ###
-'dolphin-2.9.1-llama-3-70b': dolphin_2_9_1_llama_3_70b,
-
+'openchat-3.5': openchat_3_5,
+
### x.ai ###
'grok-2': grok_2,
@@ -972,6 +808,10 @@ class ModelUtils:
### HuggingFaceH4 ###
'zephyr-7b': zephyr_7b,
+
+
+### Inferless ###
+'neural-7b': neural_7b,