Spaces:
Runtime error
Runtime error
Clémentine
commited on
Commit
·
95f85ed
1
Parent(s):
eedc0f1
added more nuance in ft models
Browse files- app.py +13 -2
- src/assets/text_content.py +6 -5
- src/auto_leaderboard/model_metadata_type.py +419 -412
app.py
CHANGED
|
@@ -328,7 +328,13 @@ with demo:
|
|
| 328 |
)
|
| 329 |
filter_columns = gr.Radio(
|
| 330 |
label="⏚ Filter model types",
|
| 331 |
-
choices = [
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 332 |
value="all",
|
| 333 |
elem_id="filter-columns"
|
| 334 |
)
|
|
@@ -404,7 +410,12 @@ with demo:
|
|
| 404 |
False, label="Private", visible=not IS_PUBLIC
|
| 405 |
)
|
| 406 |
model_type = gr.Dropdown(
|
| 407 |
-
choices=[
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 408 |
label="Model type",
|
| 409 |
multiselect=False,
|
| 410 |
value=None,
|
|
|
|
| 328 |
)
|
| 329 |
filter_columns = gr.Radio(
|
| 330 |
label="⏚ Filter model types",
|
| 331 |
+
choices = [
|
| 332 |
+
"all",
|
| 333 |
+
ModelType.PT.to_str(),
|
| 334 |
+
ModelType.FT.to_str(),
|
| 335 |
+
ModelType.IFT.to_str(),
|
| 336 |
+
ModelType.RL.to_str(),
|
| 337 |
+
],
|
| 338 |
value="all",
|
| 339 |
elem_id="filter-columns"
|
| 340 |
)
|
|
|
|
| 410 |
False, label="Private", visible=not IS_PUBLIC
|
| 411 |
)
|
| 412 |
model_type = gr.Dropdown(
|
| 413 |
+
choices=[
|
| 414 |
+
ModelType.PT.to_str(" : "),
|
| 415 |
+
ModelType.FT.to_str(" : "),
|
| 416 |
+
ModelType.IFT.to_str(" : "),
|
| 417 |
+
ModelType.RL.to_str(" : "),
|
| 418 |
+
],
|
| 419 |
label="Model type",
|
| 420 |
multiselect=False,
|
| 421 |
value=None,
|
src/assets/text_content.py
CHANGED
|
@@ -1,3 +1,5 @@
|
|
|
|
|
|
|
|
| 1 |
CHANGELOG_TEXT = f"""
|
| 2 |
## [2023-06-19]
|
| 3 |
- Added model type column
|
|
@@ -62,8 +64,6 @@ INTRODUCTION_TEXT = f"""
|
|
| 62 |
🤗 Anyone from the community can submit a model for automated evaluation on the 🤗 GPU cluster, as long as it is a 🤗 Transformers model with weights on the Hub. We also support evaluation of models with delta-weights for non-commercial licensed models, such as the original LLaMa release.
|
| 63 |
|
| 64 |
Other cool benchmarks for LLMs are developed at HuggingFace, go check them out: 🙋🤖 [human and GPT4 evals](https://huggingface.co/spaces/HuggingFaceH4/human_eval_llm_leaderboard), 🖥️ [performance benchmarks](https://huggingface.co/spaces/optimum/llm-perf-leaderboard)
|
| 65 |
-
|
| 66 |
-
🟢: Base pretrained model – 🔶: Finetuned model – 🟦: Model using RL (read more details in "About" tab)
|
| 67 |
"""
|
| 68 |
|
| 69 |
LLM_BENCHMARKS_TEXT = f"""
|
|
@@ -131,9 +131,10 @@ To get more information about quantization, see:
|
|
| 131 |
- 4 bits: [blog post](https://huggingface.co/blog/4bit-transformers-bitsandbytes), [paper](https://arxiv.org/abs/2305.14314)
|
| 132 |
|
| 133 |
### Icons
|
| 134 |
-
|
| 135 |
-
|
| 136 |
-
|
|
|
|
| 137 |
If there is no icon, we have not uploaded the information on the model yet, feel free to open an issue with the model information!
|
| 138 |
|
| 139 |
|
|
|
|
| 1 |
+
from ..auto_leaderboard.model_metadata_type import ModelType
|
| 2 |
+
|
| 3 |
CHANGELOG_TEXT = f"""
|
| 4 |
## [2023-06-19]
|
| 5 |
- Added model type column
|
|
|
|
| 64 |
🤗 Anyone from the community can submit a model for automated evaluation on the 🤗 GPU cluster, as long as it is a 🤗 Transformers model with weights on the Hub. We also support evaluation of models with delta-weights for non-commercial licensed models, such as the original LLaMa release.
|
| 65 |
|
| 66 |
Other cool benchmarks for LLMs are developed at HuggingFace, go check them out: 🙋🤖 [human and GPT4 evals](https://huggingface.co/spaces/HuggingFaceH4/human_eval_llm_leaderboard), 🖥️ [performance benchmarks](https://huggingface.co/spaces/optimum/llm-perf-leaderboard)
|
|
|
|
|
|
|
| 67 |
"""
|
| 68 |
|
| 69 |
LLM_BENCHMARKS_TEXT = f"""
|
|
|
|
| 131 |
- 4 bits: [blog post](https://huggingface.co/blog/4bit-transformers-bitsandbytes), [paper](https://arxiv.org/abs/2305.14314)
|
| 132 |
|
| 133 |
### Icons
|
| 134 |
+
{ModelType.PT.to_str(" : ")} model
|
| 135 |
+
{ModelType.FT.to_str(" : ")} model
|
| 136 |
+
{ModelType.IFT.to_str(" : ")} model
|
| 137 |
+
{ModelType.RL.to_str(" : ")} model
|
| 138 |
If there is no icon, we have not uploaded the information on the model yet, feel free to open an issue with the model information!
|
| 139 |
|
| 140 |
|
src/auto_leaderboard/model_metadata_type.py
CHANGED
|
@@ -12,132 +12,136 @@ class ModelInfo:
|
|
| 12 |
|
| 13 |
class ModelType(Enum):
|
| 14 |
PT = ModelInfo(name="pretrained", symbol="🟢")
|
| 15 |
-
|
| 16 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
|
| 18 |
|
| 19 |
TYPE_METADATA: Dict[str, ModelType] = {
|
| 20 |
-
"notstoic/PygmalionCoT-7b": ModelType.
|
| 21 |
-
"aisquared/dlite-v1-355m": ModelType.
|
| 22 |
-
"aisquared/dlite-v1-1_5b": ModelType.
|
| 23 |
-
"aisquared/dlite-v1-774m": ModelType.
|
| 24 |
-
"aisquared/dlite-v1-124m": ModelType.
|
| 25 |
-
"aisquared/chopt-2_7b": ModelType.
|
| 26 |
-
"aisquared/dlite-v2-124m": ModelType.
|
| 27 |
-
"aisquared/dlite-v2-774m": ModelType.
|
| 28 |
-
"aisquared/dlite-v2-1_5b": ModelType.
|
| 29 |
-
"aisquared/chopt-1_3b": ModelType.
|
| 30 |
-
"aisquared/dlite-v2-355m": ModelType.
|
| 31 |
-
"augtoma/qCammel-13": ModelType.
|
| 32 |
-
"Aspik101/Llama-2-7b-hf-instruct-pl-lora_unload": ModelType.
|
| 33 |
-
"Aspik101/vicuna-7b-v1.3-instruct-pl-lora_unload": ModelType.
|
| 34 |
-
"TheBloke/alpaca-lora-65B-HF": ModelType.
|
| 35 |
-
"TheBloke/tulu-7B-fp16": ModelType.
|
| 36 |
-
"TheBloke/guanaco-7B-HF": ModelType.
|
| 37 |
-
"TheBloke/koala-7B-HF": ModelType.
|
| 38 |
-
"TheBloke/wizardLM-7B-HF": ModelType.
|
| 39 |
-
"TheBloke/airoboros-13B-HF": ModelType.
|
| 40 |
-
"TheBloke/koala-13B-HF": ModelType.
|
| 41 |
-
"TheBloke/Wizard-Vicuna-7B-Uncensored-HF": ModelType.
|
| 42 |
-
"TheBloke/dromedary-65b-lora-HF": ModelType.
|
| 43 |
-
"TheBloke/wizardLM-13B-1.0-fp16": ModelType.
|
| 44 |
-
"TheBloke/WizardLM-13B-V1-1-SuperHOT-8K-fp16": ModelType.
|
| 45 |
-
"TheBloke/Wizard-Vicuna-30B-Uncensored-fp16": ModelType.
|
| 46 |
-
"TheBloke/wizard-vicuna-13B-HF": ModelType.
|
| 47 |
-
"TheBloke/UltraLM-13B-fp16": ModelType.
|
| 48 |
-
"TheBloke/OpenAssistant-
|
| 49 |
-
"TheBloke/vicuna-13B-1.1-HF": ModelType.
|
| 50 |
-
"TheBloke/guanaco-13B-HF": ModelType.
|
| 51 |
-
"TheBloke/guanaco-65B-HF": ModelType.
|
| 52 |
-
"TheBloke/airoboros-7b-gpt4-fp16": ModelType.
|
| 53 |
-
"TheBloke/llama-30b-supercot-SuperHOT-8K-fp16": ModelType.
|
| 54 |
"TheBloke/Llama-2-13B-fp16": ModelType.PT,
|
| 55 |
-
"TheBloke/llama-2-70b-Guanaco-QLoRA-fp16": ModelType.
|
| 56 |
-
"TheBloke/landmark-attention-llama7b-fp16": ModelType.
|
| 57 |
-
"TheBloke/Planner-7B-fp16": ModelType.
|
| 58 |
-
"TheBloke/Wizard-Vicuna-13B-Uncensored-HF": ModelType.
|
| 59 |
-
"TheBloke/gpt4-alpaca-lora-13B-HF": ModelType.
|
| 60 |
-
"TheBloke/gpt4-x-vicuna-13B-HF": ModelType.
|
| 61 |
-
"TheBloke/gpt4-alpaca-lora_mlp-65B-HF": ModelType.
|
| 62 |
-
"TheBloke/tulu-13B-fp16": ModelType.
|
| 63 |
-
"TheBloke/VicUnlocked-alpaca-65B-QLoRA-fp16": ModelType.
|
| 64 |
-
"TheBloke/Llama-2-70B-fp16": ModelType.
|
| 65 |
-
"TheBloke/WizardLM-30B-fp16": ModelType.
|
| 66 |
-
"TheBloke/robin-13B-v2-fp16": ModelType.
|
| 67 |
-
"TheBloke/robin-33B-v2-fp16": ModelType.
|
| 68 |
-
"TheBloke/Vicuna-13B-CoT-fp16": ModelType.
|
| 69 |
-
"TheBloke/Vicuna-33B-1-3-SuperHOT-8K-fp16": ModelType.
|
| 70 |
-
"TheBloke/Wizard-Vicuna-30B-Superhot-8K-fp16": ModelType.
|
| 71 |
-
"TheBloke/Nous-Hermes-13B-SuperHOT-8K-fp16": ModelType.
|
| 72 |
-
"TheBloke/GPlatty-30B-SuperHOT-8K-fp16": ModelType.
|
| 73 |
-
"TheBloke/CAMEL-33B-Combined-Data-SuperHOT-8K-fp16": ModelType.
|
| 74 |
-
"TheBloke/Chinese-Alpaca-33B-SuperHOT-8K-fp16": ModelType.
|
| 75 |
-
"jphme/orca_mini_v2_ger_7b": ModelType.
|
| 76 |
-
"Ejafa/vicuna_7B_vanilla_1.1": ModelType.
|
| 77 |
-
"kevinpro/Vicuna-13B-CoT": ModelType.
|
| 78 |
-
"AlekseyKorshuk/pygmalion-6b-vicuna-chatml": ModelType.
|
| 79 |
-
"AlekseyKorshuk/chatml-pyg-v1": ModelType.
|
| 80 |
-
"concedo/Vicuzard-30B-Uncensored": ModelType.
|
| 81 |
-
"concedo/OPT-19M-ChatSalad": ModelType.
|
| 82 |
-
"concedo/Pythia-70M-ChatSalad": ModelType.
|
| 83 |
-
"digitous/13B-HyperMantis": ModelType.
|
| 84 |
-
"digitous/Adventien-GPTJ": ModelType.
|
| 85 |
-
"digitous/Alpacino13b": ModelType.
|
| 86 |
-
"digitous/GPT-R": ModelType.
|
| 87 |
-
"digitous/Javelin-R": ModelType.
|
| 88 |
-
"digitous/Javalion-GPTJ": ModelType.
|
| 89 |
-
"digitous/Javalion-R": ModelType.
|
| 90 |
-
"digitous/Skegma-GPTJ": ModelType.
|
| 91 |
-
"digitous/Alpacino30b": ModelType.
|
| 92 |
-
"digitous/Janin-GPTJ": ModelType.
|
| 93 |
-
"digitous/Janin-R": ModelType.
|
| 94 |
-
"digitous/Javelin-GPTJ": ModelType.
|
| 95 |
"SaylorTwift/gpt2_test": ModelType.PT,
|
| 96 |
-
"anton-l/gpt-j-tiny-random": ModelType.
|
| 97 |
-
"Andron00e/YetAnother_Open-Llama-3B-LoRA-OpenOrca": ModelType.
|
| 98 |
-
"Lazycuber/pyg-instruct-wizardlm": ModelType.
|
| 99 |
-
"Lazycuber/Janemalion-6B": ModelType.
|
| 100 |
-
"IDEA-CCNL/Ziya-LLaMA-13B-Pretrain-v1": ModelType.
|
| 101 |
-
"IDEA-CCNL/Ziya-LLaMA-13B-v1": ModelType.
|
| 102 |
-
"dsvv-cair/alpaca-cleaned-llama-30b-bf16": ModelType.
|
| 103 |
"gpt2-medium": ModelType.PT,
|
| 104 |
-
"camel-ai/CAMEL-13B-Combined-Data": ModelType.
|
| 105 |
-
"camel-ai/CAMEL-13B-Role-Playing-Data": ModelType.
|
| 106 |
-
"camel-ai/CAMEL-33B-Combined-Data": ModelType.
|
| 107 |
-
"PygmalionAI/pygmalion-6b": ModelType.
|
| 108 |
-
"PygmalionAI/metharme-1.3b": ModelType.
|
| 109 |
-
"PygmalionAI/pygmalion-1.3b": ModelType.
|
| 110 |
-
"PygmalionAI/pygmalion-350m": ModelType.
|
| 111 |
-
"PygmalionAI/pygmalion-2.7b": ModelType.
|
| 112 |
-
"medalpaca/medalpaca-7b": ModelType.
|
| 113 |
-
"lilloukas/Platypus-30B": ModelType.
|
| 114 |
-
"lilloukas/GPlatty-30B": ModelType.
|
| 115 |
-
"mncai/chatdoctor": ModelType.
|
| 116 |
-
"chaoyi-wu/MedLLaMA_13B": ModelType.
|
| 117 |
-
"LoupGarou/WizardCoder-Guanaco-15B-V1.0": ModelType.
|
| 118 |
-
"LoupGarou/WizardCoder-Guanaco-15B-V1.1": ModelType.
|
| 119 |
-
"hakurei/instruct-12b": ModelType.
|
| 120 |
-
"hakurei/lotus-12B": ModelType.
|
| 121 |
-
"shibing624/chinese-llama-plus-13b-hf": ModelType.
|
| 122 |
-
"shibing624/chinese-alpaca-plus-7b-hf": ModelType.
|
| 123 |
-
"shibing624/chinese-alpaca-plus-13b-hf": ModelType.
|
| 124 |
-
"mosaicml/mpt-7b-instruct": ModelType.
|
| 125 |
-
"mosaicml/mpt-30b-chat": ModelType.
|
| 126 |
-
"mosaicml/mpt-7b-storywriter": ModelType.
|
| 127 |
-
"mosaicml/mpt-30b-instruct": ModelType.
|
| 128 |
-
"mosaicml/mpt-7b-chat": ModelType.
|
| 129 |
"mosaicml/mpt-30b": ModelType.PT,
|
| 130 |
-
"Corianas/111m": ModelType.
|
| 131 |
-
"Corianas/Quokka_1.3b": ModelType.
|
| 132 |
-
"Corianas/256_5epoch": ModelType.
|
| 133 |
-
"Corianas/Quokka_256m": ModelType.
|
| 134 |
-
"Corianas/Quokka_590m": ModelType.
|
| 135 |
-
"Corianas/gpt-j-6B-Dolly": ModelType.
|
| 136 |
-
"Corianas/Quokka_2.7b": ModelType.
|
| 137 |
-
"cyberagent/open-calm-7b": ModelType.
|
| 138 |
-
"Aspik101/Nous-Hermes-13b-pl-lora_unload": ModelType.
|
| 139 |
-
"THUDM/chatglm2-6b": ModelType.
|
| 140 |
-
"MetaIX/GPT4-X-Alpasta-30b": ModelType.
|
| 141 |
"NYTK/PULI-GPTrio": ModelType.PT,
|
| 142 |
"EleutherAI/pythia-1.3b": ModelType.PT,
|
| 143 |
"EleutherAI/pythia-2.8b-deduped": ModelType.PT,
|
|
@@ -164,195 +168,195 @@ TYPE_METADATA: Dict[str, ModelType] = {
|
|
| 164 |
"roneneldan/TinyStories-1M": ModelType.PT,
|
| 165 |
"roneneldan/TinyStories-8M": ModelType.PT,
|
| 166 |
"roneneldan/TinyStories-3M": ModelType.PT,
|
| 167 |
-
"jerryjalapeno/nart-100k-7b": ModelType.
|
| 168 |
-
"lmsys/vicuna-13b-v1.3": ModelType.
|
| 169 |
-
"lmsys/vicuna-7b-v1.3": ModelType.
|
| 170 |
-
"lmsys/vicuna-13b-v1.1": ModelType.
|
| 171 |
-
"lmsys/vicuna-13b-delta-v1.1": ModelType.
|
| 172 |
-
"lmsys/vicuna-7b-delta-v1.1": ModelType.
|
| 173 |
-
"abhiramtirumala/DialoGPT-sarcastic-medium": ModelType.
|
| 174 |
-
"haonan-li/bactrian-x-llama-13b-merged": ModelType.
|
| 175 |
-
"Gryphe/MythoLogic-13b": ModelType.
|
| 176 |
-
"Gryphe/MythoBoros-13b": ModelType.
|
| 177 |
-
"pillowtalks-ai/delta13b": ModelType.
|
| 178 |
-
"wannaphong/openthaigpt-0.1.0-beta-full-model_for_open_llm_leaderboard": ModelType.
|
| 179 |
"bigscience/bloom-7b1": ModelType.PT,
|
| 180 |
"bigcode/tiny_starcoder_py": ModelType.PT,
|
| 181 |
-
"bigcode/starcoderplus": ModelType.
|
| 182 |
"bigcode/gpt_bigcode-santacoder": ModelType.PT,
|
| 183 |
"bigcode/starcoder": ModelType.PT,
|
| 184 |
-
"Open-Orca/OpenOrca-Preview1-13B": ModelType.
|
| 185 |
-
"microsoft/DialoGPT-large": ModelType.
|
| 186 |
-
"microsoft/DialoGPT-small": ModelType.
|
| 187 |
-
"microsoft/DialoGPT-medium": ModelType.
|
| 188 |
-
"microsoft/CodeGPT-small-py": ModelType.
|
| 189 |
-
"Tincando/fiction_story_generator": ModelType.
|
| 190 |
-
"Pirr/pythia-13b-deduped-green_devil": ModelType.
|
| 191 |
-
"Aeala/GPT4-x-AlpacaDente2-30b": ModelType.
|
| 192 |
-
"Aeala/GPT4-x-AlpacaDente-30b": ModelType.
|
| 193 |
-
"Aeala/GPT4-x-Alpasta-13b": ModelType.
|
| 194 |
-
"Aeala/VicUnlocked-alpaca-30b": ModelType.
|
| 195 |
-
"Tap-M/Luna-AI-Llama2-Uncensored": ModelType.
|
| 196 |
-
"illuin/test-custom-llama": ModelType.
|
| 197 |
-
"dvruette/oasst-llama-13b-2-epochs": ModelType.
|
| 198 |
-
"dvruette/oasst-gpt-neox-20b-1000-steps": ModelType.
|
| 199 |
"dvruette/llama-13b-pretrained-dropout": ModelType.PT,
|
| 200 |
"dvruette/llama-13b-pretrained": ModelType.PT,
|
| 201 |
"dvruette/llama-13b-pretrained-sft-epoch-1": ModelType.PT,
|
| 202 |
"dvruette/llama-13b-pretrained-sft-do2": ModelType.PT,
|
| 203 |
-
"dvruette/oasst-gpt-neox-20b-3000-steps": ModelType.
|
| 204 |
-
"dvruette/oasst-pythia-12b-pretrained-sft": ModelType.
|
| 205 |
-
"dvruette/oasst-pythia-6.9b-4000-steps": ModelType.
|
| 206 |
-
"dvruette/gpt-neox-20b-full-precision": ModelType.
|
| 207 |
-
"dvruette/oasst-llama-13b-1000-steps": ModelType.
|
| 208 |
"openlm-research/open_llama_7b_700bt_preview": ModelType.PT,
|
| 209 |
"openlm-research/open_llama_7b": ModelType.PT,
|
| 210 |
"openlm-research/open_llama_7b_v2": ModelType.PT,
|
| 211 |
"openlm-research/open_llama_3b": ModelType.PT,
|
| 212 |
"openlm-research/open_llama_13b": ModelType.PT,
|
| 213 |
"openlm-research/open_llama_3b_v2": ModelType.PT,
|
| 214 |
-
"PocketDoc/Dans-PileOfSets-Mk1-llama-13b-merged": ModelType.
|
| 215 |
-
"GeorgiaTechResearchInstitute/galpaca-30b": ModelType.
|
| 216 |
-
"GeorgiaTechResearchInstitute/starcoder-gpteacher-code-instruct": ModelType.
|
| 217 |
-
"databricks/dolly-v2-7b": ModelType.
|
| 218 |
-
"databricks/dolly-v2-3b": ModelType.
|
| 219 |
-
"databricks/dolly-v2-12b": ModelType.
|
| 220 |
-
"Rachneet/gpt2-xl-alpaca": ModelType.
|
| 221 |
-
"Locutusque/gpt2-conversational-or-qa": ModelType.
|
| 222 |
-
"psyche/kogpt": ModelType.
|
| 223 |
-
"NbAiLab/nb-gpt-j-6B-alpaca": ModelType.
|
| 224 |
-
"Mikael110/llama-2-7b-guanaco-fp16": ModelType.
|
| 225 |
-
"Mikael110/llama-2-13b-guanaco-fp16": ModelType.
|
| 226 |
-
"Fredithefish/CrimsonPajama": ModelType.
|
| 227 |
-
"Fredithefish/RedPajama-INCITE-Chat-3B-ShareGPT-11K": ModelType.
|
| 228 |
-
"Fredithefish/ScarletPajama-3B-HF": ModelType.
|
| 229 |
-
"Fredithefish/RedPajama-INCITE-Chat-3B-Instruction-Tuning-with-GPT-4": ModelType.
|
| 230 |
-
"acrastt/RedPajama-INCITE-Chat-Instruct-3B-V1": ModelType.
|
| 231 |
-
"eachadea/vicuna-13b-1.1": ModelType.
|
| 232 |
-
"eachadea/vicuna-7b-1.1": ModelType.
|
| 233 |
-
"eachadea/vicuna-13b": ModelType.
|
| 234 |
-
"openaccess-ai-collective/wizard-mega-13b": ModelType.
|
| 235 |
-
"openaccess-ai-collective/manticore-13b": ModelType.
|
| 236 |
-
"openaccess-ai-collective/manticore-30b-chat-pyg-alpha": ModelType.
|
| 237 |
-
"openaccess-ai-collective/minotaur-13b": ModelType.
|
| 238 |
-
"openaccess-ai-collective/minotaur-13b-fixed": ModelType.
|
| 239 |
-
"openaccess-ai-collective/hippogriff-30b-chat": ModelType.
|
| 240 |
-
"openaccess-ai-collective/manticore-13b-chat-pyg": ModelType.
|
| 241 |
-
"pythainlp/wangchanglm-7.5B-sft-enth": ModelType.
|
| 242 |
-
"pythainlp/wangchanglm-7.5B-sft-en-sharded": ModelType.
|
| 243 |
-
"euclaise/gpt-neox-122m-minipile-digits": ModelType.
|
| 244 |
-
"stabilityai/StableBeluga1-Delta": ModelType.
|
| 245 |
-
"stabilityai/stablelm-tuned-alpha-7b": ModelType.
|
| 246 |
-
"stabilityai/StableBeluga2": ModelType.
|
| 247 |
-
"stabilityai/StableBeluga-13B": ModelType.
|
| 248 |
-
"stabilityai/StableBeluga-7B": ModelType.
|
| 249 |
"stabilityai/stablelm-base-alpha-7b": ModelType.PT,
|
| 250 |
"stabilityai/stablelm-base-alpha-3b": ModelType.PT,
|
| 251 |
-
"stabilityai/stablelm-tuned-alpha-3b": ModelType.
|
| 252 |
-
"alibidaran/medical_transcription_generator": ModelType.
|
| 253 |
-
"CalderaAI/30B-Lazarus": ModelType.
|
| 254 |
-
"CalderaAI/13B-BlueMethod": ModelType.
|
| 255 |
-
"CalderaAI/13B-Ouroboros": ModelType.
|
| 256 |
-
"KoboldAI/OPT-13B-Erebus": ModelType.
|
| 257 |
-
"KoboldAI/GPT-J-6B-Janeway": ModelType.
|
| 258 |
-
"KoboldAI/GPT-J-6B-Shinen": ModelType.
|
| 259 |
"KoboldAI/fairseq-dense-2.7B": ModelType.PT,
|
| 260 |
-
"KoboldAI/OPT-6B-nerys-v2": ModelType.
|
| 261 |
-
"KoboldAI/GPT-NeoX-20B-Skein": ModelType.
|
| 262 |
-
"KoboldAI/PPO_Pygway-6b-Mix": ModelType.
|
| 263 |
"KoboldAI/fairseq-dense-6.7B": ModelType.PT,
|
| 264 |
"KoboldAI/fairseq-dense-125M": ModelType.PT,
|
| 265 |
-
"KoboldAI/OPT-13B-Nerybus-Mix": ModelType.
|
| 266 |
-
"KoboldAI/OPT-2.7B-Erebus": ModelType.
|
| 267 |
-
"KoboldAI/OPT-350M-Nerys-v2": ModelType.
|
| 268 |
-
"KoboldAI/OPT-2.7B-Nerys-v2": ModelType.
|
| 269 |
-
"KoboldAI/OPT-2.7B-Nerybus-Mix": ModelType.
|
| 270 |
-
"KoboldAI/OPT-13B-Nerys-v2": ModelType.
|
| 271 |
-
"KoboldAI/GPT-NeoX-20B-Erebus": ModelType.
|
| 272 |
-
"KoboldAI/OPT-6.7B-Erebus": ModelType.
|
| 273 |
"KoboldAI/fairseq-dense-355M": ModelType.PT,
|
| 274 |
-
"KoboldAI/OPT-6.7B-Nerybus-Mix": ModelType.
|
| 275 |
-
"KoboldAI/GPT-J-6B-Adventure": ModelType.
|
| 276 |
-
"KoboldAI/OPT-350M-Erebus": ModelType.
|
| 277 |
-
"KoboldAI/GPT-J-6B-Skein": ModelType.
|
| 278 |
-
"KoboldAI/OPT-30B-Erebus": ModelType.
|
| 279 |
"klosax/pythia-160m-deduped-step92k-193bt": ModelType.PT,
|
| 280 |
"klosax/open_llama_3b_350bt_preview": ModelType.PT,
|
| 281 |
"klosax/openllama-3b-350bt": ModelType.PT,
|
| 282 |
"klosax/pythia-70m-deduped-step44k-92bt": ModelType.PT,
|
| 283 |
"klosax/open_llama_13b_600bt_preview": ModelType.PT,
|
| 284 |
"klosax/open_llama_7b_400bt_preview": ModelType.PT,
|
| 285 |
-
"kfkas/Llama-2-ko-7b-Chat": ModelType.
|
| 286 |
-
"WeOpenML/Alpaca-7B-v1": ModelType.
|
| 287 |
-
"WeOpenML/PandaLM-Alpaca-7B-v1": ModelType.
|
| 288 |
-
"TFLai/gpt2-turkish-uncased": ModelType.
|
| 289 |
-
"ehartford/WizardLM-13B-Uncensored": ModelType.
|
| 290 |
-
"ehartford/dolphin-llama-13b": ModelType.
|
| 291 |
-
"ehartford/Wizard-Vicuna-30B-Uncensored": ModelType.
|
| 292 |
-
"ehartford/WizardLM-30B-Uncensored": ModelType.
|
| 293 |
-
"ehartford/Wizard-Vicuna-13B-Uncensored": ModelType.
|
| 294 |
-
"ehartford/WizardLM-7B-Uncensored": ModelType.
|
| 295 |
-
"ehartford/based-30b": ModelType.
|
| 296 |
-
"ehartford/Wizard-Vicuna-7B-Uncensored": ModelType.
|
| 297 |
-
"wahaha1987/llama_7b_sharegpt94k_fastchat": ModelType.
|
| 298 |
-
"wahaha1987/llama_13b_sharegpt94k_fastchat": ModelType.
|
| 299 |
-
"OpenAssistant/oasst-sft-1-pythia-12b": ModelType.
|
| 300 |
-
"OpenAssistant/stablelm-7b-sft-v7-epoch-3": ModelType.
|
| 301 |
-
"OpenAssistant/oasst-sft-4-pythia-12b-epoch-3.5": ModelType.
|
| 302 |
-
"OpenAssistant/pythia-12b-sft-v8-2.5k-steps": ModelType.
|
| 303 |
-
"OpenAssistant/pythia-12b-sft-v8-7k-steps": ModelType.
|
| 304 |
-
"OpenAssistant/pythia-12b-pre-v8-12.5k-steps": ModelType.
|
| 305 |
-
"OpenAssistant/llama2-13b-orca-8k-3319": ModelType.
|
| 306 |
-
"junelee/wizard-vicuna-13b": ModelType.
|
| 307 |
"BreadAi/gpt-YA-1-1_160M": ModelType.PT,
|
| 308 |
"BreadAi/MuseCan": ModelType.PT,
|
| 309 |
"BreadAi/MusePy-1-2": ModelType.PT,
|
| 310 |
"BreadAi/DiscordPy": ModelType.PT,
|
| 311 |
"BreadAi/PM_modelV2": ModelType.PT,
|
| 312 |
"BreadAi/gpt-Youtube": ModelType.PT,
|
| 313 |
-
"BreadAi/StoryPy": ModelType.
|
| 314 |
-
"julianweng/Llama-2-7b-chat-orcah": ModelType.
|
| 315 |
-
"AGI-inc/lora_moe_7b_baseline": ModelType.
|
| 316 |
-
"AGI-inc/lora_moe_7b": ModelType.
|
| 317 |
-
"togethercomputer/GPT-NeoXT-Chat-Base-20B": ModelType.
|
| 318 |
-
"togethercomputer/RedPajama-INCITE-Chat-7B-v0.1": ModelType.
|
| 319 |
-
"togethercomputer/RedPajama-INCITE-Instruct-7B-v0.1": ModelType.
|
| 320 |
"togethercomputer/RedPajama-INCITE-7B-Base": ModelType.PT,
|
| 321 |
-
"togethercomputer/RedPajama-INCITE-7B-Instruct": ModelType.
|
| 322 |
"togethercomputer/RedPajama-INCITE-Base-3B-v1": ModelType.PT,
|
| 323 |
-
"togethercomputer/Pythia-Chat-Base-7B": ModelType.
|
| 324 |
"togethercomputer/RedPajama-INCITE-Base-7B-v0.1": ModelType.PT,
|
| 325 |
-
"togethercomputer/GPT-JT-6B-v1": ModelType.
|
| 326 |
-
"togethercomputer/GPT-JT-6B-v0": ModelType.
|
| 327 |
-
"togethercomputer/RedPajama-INCITE-Chat-3B-v1": ModelType.
|
| 328 |
-
"togethercomputer/RedPajama-INCITE-7B-Chat": ModelType.
|
| 329 |
-
"togethercomputer/RedPajama-INCITE-Instruct-3B-v1": ModelType.
|
| 330 |
-
"Writer/camel-5b-hf": ModelType.
|
| 331 |
"Writer/palmyra-base": ModelType.PT,
|
| 332 |
-
"MBZUAI/LaMini-GPT-1.5B": ModelType.
|
| 333 |
-
"MBZUAI/lamini-cerebras-111m": ModelType.
|
| 334 |
-
"MBZUAI/lamini-neo-1.3b": ModelType.
|
| 335 |
-
"MBZUAI/lamini-cerebras-1.3b": ModelType.
|
| 336 |
-
"MBZUAI/lamini-cerebras-256m": ModelType.
|
| 337 |
-
"MBZUAI/LaMini-GPT-124M": ModelType.
|
| 338 |
-
"MBZUAI/lamini-neo-125m": ModelType.
|
| 339 |
-
"TehVenom/DiffMerge-DollyGPT-Pygmalion": ModelType.
|
| 340 |
-
"TehVenom/PPO_Shygmalion-6b": ModelType.
|
| 341 |
-
"TehVenom/Dolly_Shygmalion-6b-Dev_V8P2": ModelType.
|
| 342 |
-
"TehVenom/Pygmalion_AlpacaLora-7b": ModelType.
|
| 343 |
-
"TehVenom/PPO_Pygway-V8p4_Dev-6b": ModelType.
|
| 344 |
-
"TehVenom/Dolly_Malion-6b": ModelType.
|
| 345 |
-
"TehVenom/PPO_Shygmalion-V8p4_Dev-6b": ModelType.
|
| 346 |
-
"TehVenom/ChanMalion": ModelType.
|
| 347 |
-
"TehVenom/GPT-J-Pyg_PPO-6B": ModelType.
|
| 348 |
-
"TehVenom/Pygmalion-13b-Merged": ModelType.
|
| 349 |
-
"TehVenom/Metharme-13b-Merged": ModelType.
|
| 350 |
-
"TehVenom/Dolly_Shygmalion-6b": ModelType.
|
| 351 |
-
"TehVenom/GPT-J-Pyg_PPO-6B-Dev-V8p4": ModelType.
|
| 352 |
-
"georgesung/llama2_7b_chat_uncensored": ModelType.
|
| 353 |
-
"vicgalle/gpt2-alpaca": ModelType.
|
| 354 |
-
"vicgalle/alpaca-7b": ModelType.
|
| 355 |
-
"vicgalle/gpt2-alpaca-gpt4": ModelType.
|
| 356 |
"facebook/opt-350m": ModelType.PT,
|
| 357 |
"facebook/opt-125m": ModelType.PT,
|
| 358 |
"facebook/xglm-4.5B": ModelType.PT,
|
|
@@ -364,81 +368,81 @@ TYPE_METADATA: Dict[str, ModelType] = {
|
|
| 364 |
"facebook/xglm-7.5B": ModelType.PT,
|
| 365 |
"facebook/xglm-564M": ModelType.PT,
|
| 366 |
"facebook/opt-30b": ModelType.PT,
|
| 367 |
-
"golaxy/gogpt-7b": ModelType.
|
| 368 |
-
"golaxy/gogpt2-7b": ModelType.
|
| 369 |
-
"golaxy/gogpt-7b-bloom": ModelType.
|
| 370 |
-
"golaxy/gogpt-3b-bloom": ModelType.
|
| 371 |
-
"psmathur/orca_mini_v2_7b": ModelType.
|
| 372 |
-
"psmathur/orca_mini_7b": ModelType.
|
| 373 |
-
"psmathur/orca_mini_3b": ModelType.
|
| 374 |
-
"psmathur/orca_mini_v2_13b": ModelType.
|
| 375 |
"gpt2-xl": ModelType.PT,
|
| 376 |
-
"lxe/Cerebras-GPT-2.7B-Alpaca-SP": ModelType.
|
| 377 |
-
"Monero/Manticore-13b-Chat-Pyg-Guanaco": ModelType.
|
| 378 |
-
"Monero/WizardLM-Uncensored-SuperCOT-StoryTelling-30b": ModelType.
|
| 379 |
-
"Monero/WizardLM-13b-OpenAssistant-Uncensored": ModelType.
|
| 380 |
-
"Monero/WizardLM-30B-Uncensored-Guanaco-SuperCOT-30b": ModelType.
|
| 381 |
-
"jzjiao/opt-1.3b-rlhf": ModelType.
|
| 382 |
-
"HuggingFaceH4/starchat-beta": ModelType.
|
| 383 |
-
"KnutJaegersberg/gpt-2-xl-EvolInstruct": ModelType.
|
| 384 |
-
"KnutJaegersberg/megatron-GPT-2-345m-EvolInstruct": ModelType.
|
| 385 |
-
"KnutJaegersberg/galactica-orca-wizardlm-1.3b": ModelType.
|
| 386 |
-
"openchat/openchat_8192": ModelType.
|
| 387 |
-
"openchat/openchat_v2": ModelType.
|
| 388 |
-
"openchat/openchat_v2_w": ModelType.
|
| 389 |
-
"ausboss/llama-13b-supercot": ModelType.
|
| 390 |
-
"ausboss/llama-30b-supercot": ModelType.
|
| 391 |
-
"Neko-Institute-of-Science/metharme-7b": ModelType.
|
| 392 |
-
"Neko-Institute-of-Science/pygmalion-7b": ModelType.
|
| 393 |
-
"SebastianSchramm/Cerebras-GPT-111M-instruction": ModelType.
|
| 394 |
-
"victor123/WizardLM-13B-1.0": ModelType.
|
| 395 |
-
"OpenBuddy/openbuddy-openllama-13b-v7-fp16": ModelType.
|
| 396 |
-
"OpenBuddy/openbuddy-llama2-13b-v8.1-fp16": ModelType.
|
| 397 |
-
"OpenBuddyEA/openbuddy-llama-30b-v7.1-bf16": ModelType.
|
| 398 |
"baichuan-inc/Baichuan-7B": ModelType.PT,
|
| 399 |
-
"tiiuae/falcon-40b-instruct": ModelType.
|
| 400 |
"tiiuae/falcon-40b": ModelType.PT,
|
| 401 |
"tiiuae/falcon-7b": ModelType.PT,
|
| 402 |
-
"YeungNLP/firefly-llama-13b": ModelType.
|
| 403 |
-
"YeungNLP/firefly-llama-13b-v1.2": ModelType.
|
| 404 |
-
"YeungNLP/firefly-llama2-13b": ModelType.
|
| 405 |
-
"YeungNLP/firefly-ziya-13b": ModelType.
|
| 406 |
-
"shaohang/Sparse0.5_OPT-1.3": ModelType.
|
| 407 |
-
"xzuyn/Alpacino-SuperCOT-13B": ModelType.
|
| 408 |
-
"xzuyn/MedicWizard-7B": ModelType.
|
| 409 |
-
"xDAN-AI/xDAN_13b_l2_lora": ModelType.
|
| 410 |
-
"beomi/KoAlpaca-Polyglot-5.8B": ModelType.
|
| 411 |
-
"beomi/llama-2-ko-7b": ModelType.
|
| 412 |
"Salesforce/codegen-6B-multi": ModelType.PT,
|
| 413 |
"Salesforce/codegen-16B-nl": ModelType.PT,
|
| 414 |
"Salesforce/codegen-6B-nl": ModelType.PT,
|
| 415 |
-
"ai-forever/rugpt3large_based_on_gpt2": ModelType.
|
| 416 |
"gpt2-large": ModelType.PT,
|
| 417 |
-
"frank098/orca_mini_3b_juniper": ModelType.
|
| 418 |
-
"frank098/WizardLM_13B_juniper": ModelType.
|
| 419 |
-
"FPHam/Free_Sydney_13b_HF": ModelType.
|
| 420 |
"huggingface/llama-13b": ModelType.PT,
|
| 421 |
"huggingface/llama-7b": ModelType.PT,
|
| 422 |
"huggingface/llama-65b": ModelType.PT,
|
| 423 |
"huggingface/llama-65b": ModelType.PT,
|
| 424 |
"huggingface/llama-30b": ModelType.PT,
|
| 425 |
-
"Henk717/chronoboros-33B": ModelType.
|
| 426 |
-
"jondurbin/airoboros-13b-gpt4-1.4": ModelType.
|
| 427 |
-
"jondurbin/airoboros-7b": ModelType.
|
| 428 |
-
"jondurbin/airoboros-7b-gpt4": ModelType.
|
| 429 |
-
"jondurbin/airoboros-7b-gpt4-1.1": ModelType.
|
| 430 |
-
"jondurbin/airoboros-7b-gpt4-1.2": ModelType.
|
| 431 |
-
"jondurbin/airoboros-7b-gpt4-1.3": ModelType.
|
| 432 |
-
"jondurbin/airoboros-7b-gpt4-1.4": ModelType.
|
| 433 |
-
"jondurbin/airoboros-l2-7b-gpt4-1.4.1": ModelType.
|
| 434 |
-
"jondurbin/airoboros-l2-13b-gpt4-1.4.1": ModelType.
|
| 435 |
-
"jondurbin/airoboros-l2-70b-gpt4-1.4.1": ModelType.
|
| 436 |
-
"jondurbin/airoboros-13b": ModelType.
|
| 437 |
-
"jondurbin/airoboros-33b-gpt4-1.4": ModelType.
|
| 438 |
-
"jondurbin/airoboros-33b-gpt4-1.2": ModelType.
|
| 439 |
-
"jondurbin/airoboros-65b-gpt4-1.2": ModelType.
|
| 440 |
-
"ariellee/SuperPlatty-30B": ModelType.
|
| 441 |
-
"danielhanchen/open_llama_3b_600bt_preview": ModelType.
|
| 442 |
"cerebras/Cerebras-GPT-256M": ModelType.PT,
|
| 443 |
"cerebras/Cerebras-GPT-1.3B": ModelType.PT,
|
| 444 |
"cerebras/Cerebras-GPT-13B": ModelType.PT,
|
|
@@ -446,92 +450,92 @@ TYPE_METADATA: Dict[str, ModelType] = {
|
|
| 446 |
"cerebras/Cerebras-GPT-111M": ModelType.PT,
|
| 447 |
"cerebras/Cerebras-GPT-6.7B": ModelType.PT,
|
| 448 |
"Yhyu13/oasst-rlhf-2-llama-30b-7k-steps-hf": ModelType.RL,
|
| 449 |
-
"Yhyu13/llama-30B-hf-openassitant": ModelType.
|
| 450 |
-
"NousResearch/Nous-Hermes-Llama2-13b": ModelType.
|
| 451 |
-
"NousResearch/Nous-Hermes-llama-2-7b": ModelType.
|
| 452 |
-
"NousResearch/Redmond-Puffin-13B": ModelType.
|
| 453 |
-
"NousResearch/Nous-Hermes-13b": ModelType.
|
| 454 |
-
"project-baize/baize-v2-7b": ModelType.
|
| 455 |
-
"project-baize/baize-v2-13b": ModelType.
|
| 456 |
-
"LLMs/WizardLM-13B-V1.0": ModelType.
|
| 457 |
-
"LLMs/AlpacaGPT4-7B-elina": ModelType.
|
| 458 |
-
"wenge-research/yayi-7b": ModelType.
|
| 459 |
-
"wenge-research/yayi-7b-llama2": ModelType.
|
| 460 |
-
"wenge-research/yayi-13b-llama2": ModelType.
|
| 461 |
-
"yhyhy3/open_llama_7b_v2_med_instruct": ModelType.
|
| 462 |
-
"llama-anon/instruct-13b": ModelType.
|
| 463 |
-
"huggingtweets/jerma985": ModelType.
|
| 464 |
-
"huggingtweets/gladosystem": ModelType.
|
| 465 |
-
"huggingtweets/bladeecity-jerma985": ModelType.
|
| 466 |
"huggyllama/llama-13b": ModelType.PT,
|
| 467 |
"huggyllama/llama-65b": ModelType.PT,
|
| 468 |
"FabbriSimo01/Facebook_opt_1.3b_Quantized": ModelType.PT,
|
| 469 |
-
"upstage/Llama-2-70b-instruct": ModelType.
|
| 470 |
-
"upstage/Llama-2-70b-instruct-1024": ModelType.
|
| 471 |
-
"upstage/llama-65b-instruct": ModelType.
|
| 472 |
-
"upstage/llama-30b-instruct-2048": ModelType.
|
| 473 |
-
"upstage/llama-30b-instruct": ModelType.
|
| 474 |
-
"WizardLM/WizardLM-13B-1.0": ModelType.
|
| 475 |
-
"WizardLM/WizardLM-13B-V1.1": ModelType.
|
| 476 |
-
"WizardLM/WizardLM-13B-V1.2": ModelType.
|
| 477 |
-
"WizardLM/WizardLM-30B-V1.0": ModelType.
|
| 478 |
-
"WizardLM/WizardCoder-15B-V1.0": ModelType.
|
| 479 |
"gpt2": ModelType.PT,
|
| 480 |
-
"keyfan/vicuna-chinese-replication-v1.1": ModelType.
|
| 481 |
-
"nthngdy/pythia-owt2-70m-100k": ModelType.
|
| 482 |
-
"nthngdy/pythia-owt2-70m-50k": ModelType.
|
| 483 |
-
"quantumaikr/KoreanLM-hf": ModelType.
|
| 484 |
-
"quantumaikr/open_llama_7b_hf": ModelType.
|
| 485 |
-
"quantumaikr/QuantumLM-70B-hf": ModelType.
|
| 486 |
-
"MayaPH/FinOPT-Lincoln": ModelType.
|
| 487 |
-
"MayaPH/FinOPT-Franklin": ModelType.
|
| 488 |
-
"MayaPH/GodziLLa-30B": ModelType.
|
| 489 |
-
"MayaPH/GodziLLa-30B-plus": ModelType.
|
| 490 |
-
"MayaPH/FinOPT-Washington": ModelType.
|
| 491 |
-
"ogimgio/gpt-neo-125m-neurallinguisticpioneers": ModelType.
|
| 492 |
-
"layoric/llama-2-13b-code-alpaca": ModelType.
|
| 493 |
-
"CobraMamba/mamba-gpt-3b": ModelType.
|
| 494 |
-
"CobraMamba/mamba-gpt-3b-v2": ModelType.
|
| 495 |
-
"CobraMamba/mamba-gpt-3b-v3": ModelType.
|
| 496 |
-
"timdettmers/guanaco-33b-merged": ModelType.
|
| 497 |
-
"elinas/chronos-33b": ModelType.
|
| 498 |
-
"heegyu/RedTulu-Uncensored-3B-0719": ModelType.
|
| 499 |
-
"heegyu/WizardVicuna-Uncensored-3B-0719": ModelType.
|
| 500 |
-
"heegyu/WizardVicuna-3B-0719": ModelType.
|
| 501 |
"meta-llama/Llama-2-7b-chat-hf": ModelType.RL,
|
| 502 |
"meta-llama/Llama-2-7b-hf": ModelType.PT,
|
| 503 |
"meta-llama/Llama-2-13b-chat-hf": ModelType.RL,
|
| 504 |
"meta-llama/Llama-2-13b-hf": ModelType.PT,
|
| 505 |
"meta-llama/Llama-2-70b-chat-hf": ModelType.RL,
|
| 506 |
"meta-llama/Llama-2-70b-hf": ModelType.PT,
|
| 507 |
-
"xhyi/PT_GPTNEO350_ATG": ModelType.
|
| 508 |
-
"h2oai/h2ogpt-gm-oasst1-en-1024-20b": ModelType.
|
| 509 |
-
"h2oai/h2ogpt-gm-oasst1-en-1024-open-llama-7b-preview-400bt": ModelType.
|
| 510 |
-
"h2oai/h2ogpt-oig-oasst1-512-6_9b": ModelType.
|
| 511 |
-
"h2oai/h2ogpt-oasst1-512-12b": ModelType.
|
| 512 |
-
"h2oai/h2ogpt-oig-oasst1-256-6_9b": ModelType.
|
| 513 |
-
"h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt": ModelType.
|
| 514 |
-
"h2oai/h2ogpt-oasst1-512-20b": ModelType.
|
| 515 |
-
"h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt-v2": ModelType.
|
| 516 |
-
"h2oai/h2ogpt-gm-oasst1-en-1024-12b": ModelType.
|
| 517 |
-
"h2oai/h2ogpt-gm-oasst1-multilang-1024-20b": ModelType.
|
| 518 |
-
"bofenghuang/vigogne-13b-instruct": ModelType.
|
| 519 |
-
"bofenghuang/vigogne-13b-chat": ModelType.
|
| 520 |
-
"bofenghuang/vigogne-2-7b-instruct": ModelType.
|
| 521 |
-
"bofenghuang/vigogne-7b-instruct": ModelType.
|
| 522 |
-
"bofenghuang/vigogne-7b-chat": ModelType.
|
| 523 |
-
"Vmware/open-llama-7b-v2-open-instruct": ModelType.
|
| 524 |
-
"VMware/open-llama-0.7T-7B-open-instruct-v1.1": ModelType.
|
| 525 |
-
"ewof/koishi-instruct-3b": ModelType.
|
| 526 |
-
"gywy/llama2-13b-chinese-v1": ModelType.
|
| 527 |
-
"GOAT-AI/GOAT-7B-Community": ModelType.
|
| 528 |
-
"psyche/kollama2-7b": ModelType.
|
| 529 |
-
"TheTravellingEngineer/llama2-7b-hf-guanaco": ModelType.
|
| 530 |
-
"beaugogh/pythia-1.4b-deduped-sharegpt": ModelType.
|
| 531 |
-
"augtoma/qCammel-70-x": ModelType.
|
| 532 |
-
"Lajonbot/Llama-2-7b-chat-hf-instruct-pl-lora_unload": ModelType.
|
| 533 |
-
"anhnv125/pygmalion-6b-roleplay": ModelType.
|
| 534 |
-
"64bits/LexPodLM-13B": ModelType.
|
| 535 |
}
|
| 536 |
|
| 537 |
|
|
@@ -547,9 +551,12 @@ def get_model_type(leaderboard_data: List[dict]):
|
|
| 547 |
elif model_data[AutoEvalColumn.model_type.name] == "pretrained" or any([i in model_data["model_name_for_query"] for i in ["pretrained"]]):
|
| 548 |
model_data[AutoEvalColumn.model_type.name] = ModelType.PT.value.name
|
| 549 |
model_data[AutoEvalColumn.model_type_symbol.name] = ModelType.PT.value.symbol + ("🔺" if is_delta else "")
|
|
|
|
|
|
|
|
|
|
| 550 |
elif model_data[AutoEvalColumn.model_type.name] == "finetuned" or any([i in model_data["model_name_for_query"] for i in ["finetuned", "-ft-"]]):
|
| 551 |
-
model_data[AutoEvalColumn.model_type.name] = ModelType.
|
| 552 |
-
model_data[AutoEvalColumn.model_type_symbol.name] = ModelType.
|
| 553 |
elif model_data[AutoEvalColumn.model_type.name] == "with RL" or any([i in model_data["model_name_for_query"] for i in ["-rl-", "-rlhf-"]]):
|
| 554 |
model_data[AutoEvalColumn.model_type.name] = ModelType.RL.value.name
|
| 555 |
model_data[AutoEvalColumn.model_type_symbol.name] = ModelType.RL.value.symbol + ("🔺" if is_delta else "")
|
|
|
|
| 12 |
|
| 13 |
class ModelType(Enum):
|
| 14 |
PT = ModelInfo(name="pretrained", symbol="🟢")
|
| 15 |
+
FT = ModelInfo(name="finetuned", symbol="🔶")
|
| 16 |
+
IFT = ModelInfo(name="instruction-tuned", symbol="⭕")
|
| 17 |
+
RL = ModelInfo(name="RL-tuned", symbol="🟦")
|
| 18 |
+
|
| 19 |
+
def to_str(self, separator = " "):
|
| 20 |
+
return f"{self.value.symbol}{separator}{self.value.name}"
|
| 21 |
|
| 22 |
|
| 23 |
TYPE_METADATA: Dict[str, ModelType] = {
|
| 24 |
+
"notstoic/PygmalionCoT-7b": ModelType.FT,
|
| 25 |
+
"aisquared/dlite-v1-355m": ModelType.FT,
|
| 26 |
+
"aisquared/dlite-v1-1_5b": ModelType.FT,
|
| 27 |
+
"aisquared/dlite-v1-774m": ModelType.FT,
|
| 28 |
+
"aisquared/dlite-v1-124m": ModelType.FT,
|
| 29 |
+
"aisquared/chopt-2_7b": ModelType.FT,
|
| 30 |
+
"aisquared/dlite-v2-124m": ModelType.FT,
|
| 31 |
+
"aisquared/dlite-v2-774m": ModelType.FT,
|
| 32 |
+
"aisquared/dlite-v2-1_5b": ModelType.FT,
|
| 33 |
+
"aisquared/chopt-1_3b": ModelType.FT,
|
| 34 |
+
"aisquared/dlite-v2-355m": ModelType.FT,
|
| 35 |
+
"augtoma/qCammel-13": ModelType.FT,
|
| 36 |
+
"Aspik101/Llama-2-7b-hf-instruct-pl-lora_unload": ModelType.FT,
|
| 37 |
+
"Aspik101/vicuna-7b-v1.3-instruct-pl-lora_unload": ModelType.FT,
|
| 38 |
+
"TheBloke/alpaca-lora-65B-HF": ModelType.FT,
|
| 39 |
+
"TheBloke/tulu-7B-fp16": ModelType.FT,
|
| 40 |
+
"TheBloke/guanaco-7B-HF": ModelType.FT,
|
| 41 |
+
"TheBloke/koala-7B-HF": ModelType.FT,
|
| 42 |
+
"TheBloke/wizardLM-7B-HF": ModelType.FT,
|
| 43 |
+
"TheBloke/airoboros-13B-HF": ModelType.FT,
|
| 44 |
+
"TheBloke/koala-13B-HF": ModelType.FT,
|
| 45 |
+
"TheBloke/Wizard-Vicuna-7B-Uncensored-HF": ModelType.FT,
|
| 46 |
+
"TheBloke/dromedary-65b-lora-HF": ModelType.FT,
|
| 47 |
+
"TheBloke/wizardLM-13B-1.0-fp16": ModelType.FT,
|
| 48 |
+
"TheBloke/WizardLM-13B-V1-1-SuperHOT-8K-fp16": ModelType.FT,
|
| 49 |
+
"TheBloke/Wizard-Vicuna-30B-Uncensored-fp16": ModelType.FT,
|
| 50 |
+
"TheBloke/wizard-vicuna-13B-HF": ModelType.FT,
|
| 51 |
+
"TheBloke/UltraLM-13B-fp16": ModelType.FT,
|
| 52 |
+
"TheBloke/OpenAssistant-FT-7-Llama-30B-HF": ModelType.FT,
|
| 53 |
+
"TheBloke/vicuna-13B-1.1-HF": ModelType.FT,
|
| 54 |
+
"TheBloke/guanaco-13B-HF": ModelType.FT,
|
| 55 |
+
"TheBloke/guanaco-65B-HF": ModelType.FT,
|
| 56 |
+
"TheBloke/airoboros-7b-gpt4-fp16": ModelType.FT,
|
| 57 |
+
"TheBloke/llama-30b-supercot-SuperHOT-8K-fp16": ModelType.FT,
|
| 58 |
"TheBloke/Llama-2-13B-fp16": ModelType.PT,
|
| 59 |
+
"TheBloke/llama-2-70b-Guanaco-QLoRA-fp16": ModelType.FT,
|
| 60 |
+
"TheBloke/landmark-attention-llama7b-fp16": ModelType.FT,
|
| 61 |
+
"TheBloke/Planner-7B-fp16": ModelType.FT,
|
| 62 |
+
"TheBloke/Wizard-Vicuna-13B-Uncensored-HF": ModelType.FT,
|
| 63 |
+
"TheBloke/gpt4-alpaca-lora-13B-HF": ModelType.FT,
|
| 64 |
+
"TheBloke/gpt4-x-vicuna-13B-HF": ModelType.FT,
|
| 65 |
+
"TheBloke/gpt4-alpaca-lora_mlp-65B-HF": ModelType.FT,
|
| 66 |
+
"TheBloke/tulu-13B-fp16": ModelType.FT,
|
| 67 |
+
"TheBloke/VicUnlocked-alpaca-65B-QLoRA-fp16": ModelType.FT,
|
| 68 |
+
"TheBloke/Llama-2-70B-fp16": ModelType.FT,
|
| 69 |
+
"TheBloke/WizardLM-30B-fp16": ModelType.FT,
|
| 70 |
+
"TheBloke/robin-13B-v2-fp16": ModelType.FT,
|
| 71 |
+
"TheBloke/robin-33B-v2-fp16": ModelType.FT,
|
| 72 |
+
"TheBloke/Vicuna-13B-CoT-fp16": ModelType.FT,
|
| 73 |
+
"TheBloke/Vicuna-33B-1-3-SuperHOT-8K-fp16": ModelType.FT,
|
| 74 |
+
"TheBloke/Wizard-Vicuna-30B-Superhot-8K-fp16": ModelType.FT,
|
| 75 |
+
"TheBloke/Nous-Hermes-13B-SuperHOT-8K-fp16": ModelType.FT,
|
| 76 |
+
"TheBloke/GPlatty-30B-SuperHOT-8K-fp16": ModelType.FT,
|
| 77 |
+
"TheBloke/CAMEL-33B-Combined-Data-SuperHOT-8K-fp16": ModelType.FT,
|
| 78 |
+
"TheBloke/Chinese-Alpaca-33B-SuperHOT-8K-fp16": ModelType.FT,
|
| 79 |
+
"jphme/orca_mini_v2_ger_7b": ModelType.FT,
|
| 80 |
+
"Ejafa/vicuna_7B_vanilla_1.1": ModelType.FT,
|
| 81 |
+
"kevinpro/Vicuna-13B-CoT": ModelType.FT,
|
| 82 |
+
"AlekseyKorshuk/pygmalion-6b-vicuna-chatml": ModelType.FT,
|
| 83 |
+
"AlekseyKorshuk/chatml-pyg-v1": ModelType.FT,
|
| 84 |
+
"concedo/Vicuzard-30B-Uncensored": ModelType.FT,
|
| 85 |
+
"concedo/OPT-19M-ChatSalad": ModelType.FT,
|
| 86 |
+
"concedo/Pythia-70M-ChatSalad": ModelType.FT,
|
| 87 |
+
"digitous/13B-HyperMantis": ModelType.FT,
|
| 88 |
+
"digitous/Adventien-GPTJ": ModelType.FT,
|
| 89 |
+
"digitous/Alpacino13b": ModelType.FT,
|
| 90 |
+
"digitous/GPT-R": ModelType.FT,
|
| 91 |
+
"digitous/Javelin-R": ModelType.FT,
|
| 92 |
+
"digitous/Javalion-GPTJ": ModelType.FT,
|
| 93 |
+
"digitous/Javalion-R": ModelType.FT,
|
| 94 |
+
"digitous/Skegma-GPTJ": ModelType.FT,
|
| 95 |
+
"digitous/Alpacino30b": ModelType.FT,
|
| 96 |
+
"digitous/Janin-GPTJ": ModelType.FT,
|
| 97 |
+
"digitous/Janin-R": ModelType.FT,
|
| 98 |
+
"digitous/Javelin-GPTJ": ModelType.FT,
|
| 99 |
"SaylorTwift/gpt2_test": ModelType.PT,
|
| 100 |
+
"anton-l/gpt-j-tiny-random": ModelType.FT,
|
| 101 |
+
"Andron00e/YetAnother_Open-Llama-3B-LoRA-OpenOrca": ModelType.FT,
|
| 102 |
+
"Lazycuber/pyg-instruct-wizardlm": ModelType.FT,
|
| 103 |
+
"Lazycuber/Janemalion-6B": ModelType.FT,
|
| 104 |
+
"IDEA-CCNL/Ziya-LLaMA-13B-Pretrain-v1": ModelType.FT,
|
| 105 |
+
"IDEA-CCNL/Ziya-LLaMA-13B-v1": ModelType.FT,
|
| 106 |
+
"dsvv-cair/alpaca-cleaned-llama-30b-bf16": ModelType.FT,
|
| 107 |
"gpt2-medium": ModelType.PT,
|
| 108 |
+
"camel-ai/CAMEL-13B-Combined-Data": ModelType.FT,
|
| 109 |
+
"camel-ai/CAMEL-13B-Role-Playing-Data": ModelType.FT,
|
| 110 |
+
"camel-ai/CAMEL-33B-Combined-Data": ModelType.FT,
|
| 111 |
+
"PygmalionAI/pygmalion-6b": ModelType.FT,
|
| 112 |
+
"PygmalionAI/metharme-1.3b": ModelType.FT,
|
| 113 |
+
"PygmalionAI/pygmalion-1.3b": ModelType.FT,
|
| 114 |
+
"PygmalionAI/pygmalion-350m": ModelType.FT,
|
| 115 |
+
"PygmalionAI/pygmalion-2.7b": ModelType.FT,
|
| 116 |
+
"medalpaca/medalpaca-7b": ModelType.FT,
|
| 117 |
+
"lilloukas/Platypus-30B": ModelType.FT,
|
| 118 |
+
"lilloukas/GPlatty-30B": ModelType.FT,
|
| 119 |
+
"mncai/chatdoctor": ModelType.FT,
|
| 120 |
+
"chaoyi-wu/MedLLaMA_13B": ModelType.FT,
|
| 121 |
+
"LoupGarou/WizardCoder-Guanaco-15B-V1.0": ModelType.FT,
|
| 122 |
+
"LoupGarou/WizardCoder-Guanaco-15B-V1.1": ModelType.FT,
|
| 123 |
+
"hakurei/instruct-12b": ModelType.FT,
|
| 124 |
+
"hakurei/lotus-12B": ModelType.FT,
|
| 125 |
+
"shibing624/chinese-llama-plus-13b-hf": ModelType.FT,
|
| 126 |
+
"shibing624/chinese-alpaca-plus-7b-hf": ModelType.FT,
|
| 127 |
+
"shibing624/chinese-alpaca-plus-13b-hf": ModelType.FT,
|
| 128 |
+
"mosaicml/mpt-7b-instruct": ModelType.FT,
|
| 129 |
+
"mosaicml/mpt-30b-chat": ModelType.FT,
|
| 130 |
+
"mosaicml/mpt-7b-storywriter": ModelType.FT,
|
| 131 |
+
"mosaicml/mpt-30b-instruct": ModelType.FT,
|
| 132 |
+
"mosaicml/mpt-7b-chat": ModelType.FT,
|
| 133 |
"mosaicml/mpt-30b": ModelType.PT,
|
| 134 |
+
"Corianas/111m": ModelType.FT,
|
| 135 |
+
"Corianas/Quokka_1.3b": ModelType.FT,
|
| 136 |
+
"Corianas/256_5epoch": ModelType.FT,
|
| 137 |
+
"Corianas/Quokka_256m": ModelType.FT,
|
| 138 |
+
"Corianas/Quokka_590m": ModelType.FT,
|
| 139 |
+
"Corianas/gpt-j-6B-Dolly": ModelType.FT,
|
| 140 |
+
"Corianas/Quokka_2.7b": ModelType.FT,
|
| 141 |
+
"cyberagent/open-calm-7b": ModelType.FT,
|
| 142 |
+
"Aspik101/Nous-Hermes-13b-pl-lora_unload": ModelType.FT,
|
| 143 |
+
"THUDM/chatglm2-6b": ModelType.FT,
|
| 144 |
+
"MetaIX/GPT4-X-Alpasta-30b": ModelType.FT,
|
| 145 |
"NYTK/PULI-GPTrio": ModelType.PT,
|
| 146 |
"EleutherAI/pythia-1.3b": ModelType.PT,
|
| 147 |
"EleutherAI/pythia-2.8b-deduped": ModelType.PT,
|
|
|
|
| 168 |
"roneneldan/TinyStories-1M": ModelType.PT,
|
| 169 |
"roneneldan/TinyStories-8M": ModelType.PT,
|
| 170 |
"roneneldan/TinyStories-3M": ModelType.PT,
|
| 171 |
+
"jerryjalapeno/nart-100k-7b": ModelType.FT,
|
| 172 |
+
"lmsys/vicuna-13b-v1.3": ModelType.FT,
|
| 173 |
+
"lmsys/vicuna-7b-v1.3": ModelType.FT,
|
| 174 |
+
"lmsys/vicuna-13b-v1.1": ModelType.FT,
|
| 175 |
+
"lmsys/vicuna-13b-delta-v1.1": ModelType.FT,
|
| 176 |
+
"lmsys/vicuna-7b-delta-v1.1": ModelType.FT,
|
| 177 |
+
"abhiramtirumala/DialoGPT-sarcastic-medium": ModelType.FT,
|
| 178 |
+
"haonan-li/bactrian-x-llama-13b-merged": ModelType.FT,
|
| 179 |
+
"Gryphe/MythoLogic-13b": ModelType.FT,
|
| 180 |
+
"Gryphe/MythoBoros-13b": ModelType.FT,
|
| 181 |
+
"pillowtalks-ai/delta13b": ModelType.FT,
|
| 182 |
+
"wannaphong/openthaigpt-0.1.0-beta-full-model_for_open_llm_leaderboard": ModelType.FT,
|
| 183 |
"bigscience/bloom-7b1": ModelType.PT,
|
| 184 |
"bigcode/tiny_starcoder_py": ModelType.PT,
|
| 185 |
+
"bigcode/starcoderplus": ModelType.FT,
|
| 186 |
"bigcode/gpt_bigcode-santacoder": ModelType.PT,
|
| 187 |
"bigcode/starcoder": ModelType.PT,
|
| 188 |
+
"Open-Orca/OpenOrca-Preview1-13B": ModelType.FT,
|
| 189 |
+
"microsoft/DialoGPT-large": ModelType.FT,
|
| 190 |
+
"microsoft/DialoGPT-small": ModelType.FT,
|
| 191 |
+
"microsoft/DialoGPT-medium": ModelType.FT,
|
| 192 |
+
"microsoft/CodeGPT-small-py": ModelType.FT,
|
| 193 |
+
"Tincando/fiction_story_generator": ModelType.FT,
|
| 194 |
+
"Pirr/pythia-13b-deduped-green_devil": ModelType.FT,
|
| 195 |
+
"Aeala/GPT4-x-AlpacaDente2-30b": ModelType.FT,
|
| 196 |
+
"Aeala/GPT4-x-AlpacaDente-30b": ModelType.FT,
|
| 197 |
+
"Aeala/GPT4-x-Alpasta-13b": ModelType.FT,
|
| 198 |
+
"Aeala/VicUnlocked-alpaca-30b": ModelType.FT,
|
| 199 |
+
"Tap-M/Luna-AI-Llama2-Uncensored": ModelType.FT,
|
| 200 |
+
"illuin/test-custom-llama": ModelType.FT,
|
| 201 |
+
"dvruette/oasst-llama-13b-2-epochs": ModelType.FT,
|
| 202 |
+
"dvruette/oasst-gpt-neox-20b-1000-steps": ModelType.FT,
|
| 203 |
"dvruette/llama-13b-pretrained-dropout": ModelType.PT,
|
| 204 |
"dvruette/llama-13b-pretrained": ModelType.PT,
|
| 205 |
"dvruette/llama-13b-pretrained-sft-epoch-1": ModelType.PT,
|
| 206 |
"dvruette/llama-13b-pretrained-sft-do2": ModelType.PT,
|
| 207 |
+
"dvruette/oasst-gpt-neox-20b-3000-steps": ModelType.FT,
|
| 208 |
+
"dvruette/oasst-pythia-12b-pretrained-sft": ModelType.PT,
|
| 209 |
+
"dvruette/oasst-pythia-6.9b-4000-steps": ModelType.FT,
|
| 210 |
+
"dvruette/gpt-neox-20b-full-precision": ModelType.FT,
|
| 211 |
+
"dvruette/oasst-llama-13b-1000-steps": ModelType.FT,
|
| 212 |
"openlm-research/open_llama_7b_700bt_preview": ModelType.PT,
|
| 213 |
"openlm-research/open_llama_7b": ModelType.PT,
|
| 214 |
"openlm-research/open_llama_7b_v2": ModelType.PT,
|
| 215 |
"openlm-research/open_llama_3b": ModelType.PT,
|
| 216 |
"openlm-research/open_llama_13b": ModelType.PT,
|
| 217 |
"openlm-research/open_llama_3b_v2": ModelType.PT,
|
| 218 |
+
"PocketDoc/Dans-PileOfSets-Mk1-llama-13b-merged": ModelType.FT,
|
| 219 |
+
"GeorgiaTechResearchInstitute/galpaca-30b": ModelType.FT,
|
| 220 |
+
"GeorgiaTechResearchInstitute/starcoder-gpteacher-code-instruct": ModelType.FT,
|
| 221 |
+
"databricks/dolly-v2-7b": ModelType.FT,
|
| 222 |
+
"databricks/dolly-v2-3b": ModelType.FT,
|
| 223 |
+
"databricks/dolly-v2-12b": ModelType.FT,
|
| 224 |
+
"Rachneet/gpt2-xl-alpaca": ModelType.FT,
|
| 225 |
+
"Locutusque/gpt2-conversational-or-qa": ModelType.FT,
|
| 226 |
+
"psyche/kogpt": ModelType.FT,
|
| 227 |
+
"NbAiLab/nb-gpt-j-6B-alpaca": ModelType.FT,
|
| 228 |
+
"Mikael110/llama-2-7b-guanaco-fp16": ModelType.FT,
|
| 229 |
+
"Mikael110/llama-2-13b-guanaco-fp16": ModelType.FT,
|
| 230 |
+
"Fredithefish/CrimsonPajama": ModelType.FT,
|
| 231 |
+
"Fredithefish/RedPajama-INCITE-Chat-3B-ShareGPT-11K": ModelType.FT,
|
| 232 |
+
"Fredithefish/ScarletPajama-3B-HF": ModelType.FT,
|
| 233 |
+
"Fredithefish/RedPajama-INCITE-Chat-3B-Instruction-Tuning-with-GPT-4": ModelType.FT,
|
| 234 |
+
"acrastt/RedPajama-INCITE-Chat-Instruct-3B-V1": ModelType.FT,
|
| 235 |
+
"eachadea/vicuna-13b-1.1": ModelType.FT,
|
| 236 |
+
"eachadea/vicuna-7b-1.1": ModelType.FT,
|
| 237 |
+
"eachadea/vicuna-13b": ModelType.FT,
|
| 238 |
+
"openaccess-ai-collective/wizard-mega-13b": ModelType.FT,
|
| 239 |
+
"openaccess-ai-collective/manticore-13b": ModelType.FT,
|
| 240 |
+
"openaccess-ai-collective/manticore-30b-chat-pyg-alpha": ModelType.FT,
|
| 241 |
+
"openaccess-ai-collective/minotaur-13b": ModelType.FT,
|
| 242 |
+
"openaccess-ai-collective/minotaur-13b-fixed": ModelType.FT,
|
| 243 |
+
"openaccess-ai-collective/hippogriff-30b-chat": ModelType.FT,
|
| 244 |
+
"openaccess-ai-collective/manticore-13b-chat-pyg": ModelType.FT,
|
| 245 |
+
"pythainlp/wangchanglm-7.5B-sft-enth": ModelType.FT,
|
| 246 |
+
"pythainlp/wangchanglm-7.5B-sft-en-sharded": ModelType.FT,
|
| 247 |
+
"euclaise/gpt-neox-122m-minipile-digits": ModelType.FT,
|
| 248 |
+
"stabilityai/StableBeluga1-Delta": ModelType.FT,
|
| 249 |
+
"stabilityai/stablelm-tuned-alpha-7b": ModelType.FT,
|
| 250 |
+
"stabilityai/StableBeluga2": ModelType.FT,
|
| 251 |
+
"stabilityai/StableBeluga-13B": ModelType.FT,
|
| 252 |
+
"stabilityai/StableBeluga-7B": ModelType.FT,
|
| 253 |
"stabilityai/stablelm-base-alpha-7b": ModelType.PT,
|
| 254 |
"stabilityai/stablelm-base-alpha-3b": ModelType.PT,
|
| 255 |
+
"stabilityai/stablelm-tuned-alpha-3b": ModelType.FT,
|
| 256 |
+
"alibidaran/medical_transcription_generator": ModelType.FT,
|
| 257 |
+
"CalderaAI/30B-Lazarus": ModelType.FT,
|
| 258 |
+
"CalderaAI/13B-BlueMethod": ModelType.FT,
|
| 259 |
+
"CalderaAI/13B-Ouroboros": ModelType.FT,
|
| 260 |
+
"KoboldAI/OPT-13B-Erebus": ModelType.FT,
|
| 261 |
+
"KoboldAI/GPT-J-6B-Janeway": ModelType.FT,
|
| 262 |
+
"KoboldAI/GPT-J-6B-Shinen": ModelType.FT,
|
| 263 |
"KoboldAI/fairseq-dense-2.7B": ModelType.PT,
|
| 264 |
+
"KoboldAI/OPT-6B-nerys-v2": ModelType.FT,
|
| 265 |
+
"KoboldAI/GPT-NeoX-20B-Skein": ModelType.FT,
|
| 266 |
+
"KoboldAI/PPO_Pygway-6b-Mix": ModelType.FT,
|
| 267 |
"KoboldAI/fairseq-dense-6.7B": ModelType.PT,
|
| 268 |
"KoboldAI/fairseq-dense-125M": ModelType.PT,
|
| 269 |
+
"KoboldAI/OPT-13B-Nerybus-Mix": ModelType.FT,
|
| 270 |
+
"KoboldAI/OPT-2.7B-Erebus": ModelType.FT,
|
| 271 |
+
"KoboldAI/OPT-350M-Nerys-v2": ModelType.FT,
|
| 272 |
+
"KoboldAI/OPT-2.7B-Nerys-v2": ModelType.FT,
|
| 273 |
+
"KoboldAI/OPT-2.7B-Nerybus-Mix": ModelType.FT,
|
| 274 |
+
"KoboldAI/OPT-13B-Nerys-v2": ModelType.FT,
|
| 275 |
+
"KoboldAI/GPT-NeoX-20B-Erebus": ModelType.FT,
|
| 276 |
+
"KoboldAI/OPT-6.7B-Erebus": ModelType.FT,
|
| 277 |
"KoboldAI/fairseq-dense-355M": ModelType.PT,
|
| 278 |
+
"KoboldAI/OPT-6.7B-Nerybus-Mix": ModelType.FT,
|
| 279 |
+
"KoboldAI/GPT-J-6B-Adventure": ModelType.FT,
|
| 280 |
+
"KoboldAI/OPT-350M-Erebus": ModelType.FT,
|
| 281 |
+
"KoboldAI/GPT-J-6B-Skein": ModelType.FT,
|
| 282 |
+
"KoboldAI/OPT-30B-Erebus": ModelType.FT,
|
| 283 |
"klosax/pythia-160m-deduped-step92k-193bt": ModelType.PT,
|
| 284 |
"klosax/open_llama_3b_350bt_preview": ModelType.PT,
|
| 285 |
"klosax/openllama-3b-350bt": ModelType.PT,
|
| 286 |
"klosax/pythia-70m-deduped-step44k-92bt": ModelType.PT,
|
| 287 |
"klosax/open_llama_13b_600bt_preview": ModelType.PT,
|
| 288 |
"klosax/open_llama_7b_400bt_preview": ModelType.PT,
|
| 289 |
+
"kfkas/Llama-2-ko-7b-Chat": ModelType.FT,
|
| 290 |
+
"WeOpenML/Alpaca-7B-v1": ModelType.FT,
|
| 291 |
+
"WeOpenML/PandaLM-Alpaca-7B-v1": ModelType.FT,
|
| 292 |
+
"TFLai/gpt2-turkish-uncased": ModelType.FT,
|
| 293 |
+
"ehartford/WizardLM-13B-Uncensored": ModelType.FT,
|
| 294 |
+
"ehartford/dolphin-llama-13b": ModelType.FT,
|
| 295 |
+
"ehartford/Wizard-Vicuna-30B-Uncensored": ModelType.FT,
|
| 296 |
+
"ehartford/WizardLM-30B-Uncensored": ModelType.FT,
|
| 297 |
+
"ehartford/Wizard-Vicuna-13B-Uncensored": ModelType.FT,
|
| 298 |
+
"ehartford/WizardLM-7B-Uncensored": ModelType.FT,
|
| 299 |
+
"ehartford/based-30b": ModelType.FT,
|
| 300 |
+
"ehartford/Wizard-Vicuna-7B-Uncensored": ModelType.FT,
|
| 301 |
+
"wahaha1987/llama_7b_sharegpt94k_fastchat": ModelType.FT,
|
| 302 |
+
"wahaha1987/llama_13b_sharegpt94k_fastchat": ModelType.FT,
|
| 303 |
+
"OpenAssistant/oasst-sft-1-pythia-12b": ModelType.FT,
|
| 304 |
+
"OpenAssistant/stablelm-7b-sft-v7-epoch-3": ModelType.FT,
|
| 305 |
+
"OpenAssistant/oasst-sft-4-pythia-12b-epoch-3.5": ModelType.FT,
|
| 306 |
+
"OpenAssistant/pythia-12b-sft-v8-2.5k-steps": ModelType.FT,
|
| 307 |
+
"OpenAssistant/pythia-12b-sft-v8-7k-steps": ModelType.FT,
|
| 308 |
+
"OpenAssistant/pythia-12b-pre-v8-12.5k-steps": ModelType.FT,
|
| 309 |
+
"OpenAssistant/llama2-13b-orca-8k-3319": ModelType.FT,
|
| 310 |
+
"junelee/wizard-vicuna-13b": ModelType.FT,
|
| 311 |
"BreadAi/gpt-YA-1-1_160M": ModelType.PT,
|
| 312 |
"BreadAi/MuseCan": ModelType.PT,
|
| 313 |
"BreadAi/MusePy-1-2": ModelType.PT,
|
| 314 |
"BreadAi/DiscordPy": ModelType.PT,
|
| 315 |
"BreadAi/PM_modelV2": ModelType.PT,
|
| 316 |
"BreadAi/gpt-Youtube": ModelType.PT,
|
| 317 |
+
"BreadAi/StoryPy": ModelType.FT,
|
| 318 |
+
"julianweng/Llama-2-7b-chat-orcah": ModelType.FT,
|
| 319 |
+
"AGI-inc/lora_moe_7b_baseline": ModelType.FT,
|
| 320 |
+
"AGI-inc/lora_moe_7b": ModelType.FT,
|
| 321 |
+
"togethercomputer/GPT-NeoXT-Chat-Base-20B": ModelType.FT,
|
| 322 |
+
"togethercomputer/RedPajama-INCITE-Chat-7B-v0.1": ModelType.FT,
|
| 323 |
+
"togethercomputer/RedPajama-INCITE-Instruct-7B-v0.1": ModelType.FT,
|
| 324 |
"togethercomputer/RedPajama-INCITE-7B-Base": ModelType.PT,
|
| 325 |
+
"togethercomputer/RedPajama-INCITE-7B-Instruct": ModelType.FT,
|
| 326 |
"togethercomputer/RedPajama-INCITE-Base-3B-v1": ModelType.PT,
|
| 327 |
+
"togethercomputer/Pythia-Chat-Base-7B": ModelType.FT,
|
| 328 |
"togethercomputer/RedPajama-INCITE-Base-7B-v0.1": ModelType.PT,
|
| 329 |
+
"togethercomputer/GPT-JT-6B-v1": ModelType.FT,
|
| 330 |
+
"togethercomputer/GPT-JT-6B-v0": ModelType.FT,
|
| 331 |
+
"togethercomputer/RedPajama-INCITE-Chat-3B-v1": ModelType.FT,
|
| 332 |
+
"togethercomputer/RedPajama-INCITE-7B-Chat": ModelType.FT,
|
| 333 |
+
"togethercomputer/RedPajama-INCITE-Instruct-3B-v1": ModelType.FT,
|
| 334 |
+
"Writer/camel-5b-hf": ModelType.FT,
|
| 335 |
"Writer/palmyra-base": ModelType.PT,
|
| 336 |
+
"MBZUAI/LaMini-GPT-1.5B": ModelType.FT,
|
| 337 |
+
"MBZUAI/lamini-cerebras-111m": ModelType.FT,
|
| 338 |
+
"MBZUAI/lamini-neo-1.3b": ModelType.FT,
|
| 339 |
+
"MBZUAI/lamini-cerebras-1.3b": ModelType.FT,
|
| 340 |
+
"MBZUAI/lamini-cerebras-256m": ModelType.FT,
|
| 341 |
+
"MBZUAI/LaMini-GPT-124M": ModelType.FT,
|
| 342 |
+
"MBZUAI/lamini-neo-125m": ModelType.FT,
|
| 343 |
+
"TehVenom/DiffMerge-DollyGPT-Pygmalion": ModelType.FT,
|
| 344 |
+
"TehVenom/PPO_Shygmalion-6b": ModelType.FT,
|
| 345 |
+
"TehVenom/Dolly_Shygmalion-6b-Dev_V8P2": ModelType.FT,
|
| 346 |
+
"TehVenom/Pygmalion_AlpacaLora-7b": ModelType.FT,
|
| 347 |
+
"TehVenom/PPO_Pygway-V8p4_Dev-6b": ModelType.FT,
|
| 348 |
+
"TehVenom/Dolly_Malion-6b": ModelType.FT,
|
| 349 |
+
"TehVenom/PPO_Shygmalion-V8p4_Dev-6b": ModelType.FT,
|
| 350 |
+
"TehVenom/ChanMalion": ModelType.FT,
|
| 351 |
+
"TehVenom/GPT-J-Pyg_PPO-6B": ModelType.FT,
|
| 352 |
+
"TehVenom/Pygmalion-13b-Merged": ModelType.FT,
|
| 353 |
+
"TehVenom/Metharme-13b-Merged": ModelType.FT,
|
| 354 |
+
"TehVenom/Dolly_Shygmalion-6b": ModelType.FT,
|
| 355 |
+
"TehVenom/GPT-J-Pyg_PPO-6B-Dev-V8p4": ModelType.FT,
|
| 356 |
+
"georgesung/llama2_7b_chat_uncensored": ModelType.FT,
|
| 357 |
+
"vicgalle/gpt2-alpaca": ModelType.FT,
|
| 358 |
+
"vicgalle/alpaca-7b": ModelType.FT,
|
| 359 |
+
"vicgalle/gpt2-alpaca-gpt4": ModelType.FT,
|
| 360 |
"facebook/opt-350m": ModelType.PT,
|
| 361 |
"facebook/opt-125m": ModelType.PT,
|
| 362 |
"facebook/xglm-4.5B": ModelType.PT,
|
|
|
|
| 368 |
"facebook/xglm-7.5B": ModelType.PT,
|
| 369 |
"facebook/xglm-564M": ModelType.PT,
|
| 370 |
"facebook/opt-30b": ModelType.PT,
|
| 371 |
+
"golaxy/gogpt-7b": ModelType.FT,
|
| 372 |
+
"golaxy/gogpt2-7b": ModelType.FT,
|
| 373 |
+
"golaxy/gogpt-7b-bloom": ModelType.FT,
|
| 374 |
+
"golaxy/gogpt-3b-bloom": ModelType.FT,
|
| 375 |
+
"psmathur/orca_mini_v2_7b": ModelType.FT,
|
| 376 |
+
"psmathur/orca_mini_7b": ModelType.FT,
|
| 377 |
+
"psmathur/orca_mini_3b": ModelType.FT,
|
| 378 |
+
"psmathur/orca_mini_v2_13b": ModelType.FT,
|
| 379 |
"gpt2-xl": ModelType.PT,
|
| 380 |
+
"lxe/Cerebras-GPT-2.7B-Alpaca-SP": ModelType.FT,
|
| 381 |
+
"Monero/Manticore-13b-Chat-Pyg-Guanaco": ModelType.FT,
|
| 382 |
+
"Monero/WizardLM-Uncensored-SuperCOT-StoryTelling-30b": ModelType.FT,
|
| 383 |
+
"Monero/WizardLM-13b-OpenAssistant-Uncensored": ModelType.FT,
|
| 384 |
+
"Monero/WizardLM-30B-Uncensored-Guanaco-SuperCOT-30b": ModelType.FT,
|
| 385 |
+
"jzjiao/opt-1.3b-rlhf": ModelType.FT,
|
| 386 |
+
"HuggingFaceH4/starchat-beta": ModelType.FT,
|
| 387 |
+
"KnutJaegersberg/gpt-2-xl-EvolInstruct": ModelType.FT,
|
| 388 |
+
"KnutJaegersberg/megatron-GPT-2-345m-EvolInstruct": ModelType.FT,
|
| 389 |
+
"KnutJaegersberg/galactica-orca-wizardlm-1.3b": ModelType.FT,
|
| 390 |
+
"openchat/openchat_8192": ModelType.FT,
|
| 391 |
+
"openchat/openchat_v2": ModelType.FT,
|
| 392 |
+
"openchat/openchat_v2_w": ModelType.FT,
|
| 393 |
+
"ausboss/llama-13b-supercot": ModelType.FT,
|
| 394 |
+
"ausboss/llama-30b-supercot": ModelType.FT,
|
| 395 |
+
"Neko-Institute-of-Science/metharme-7b": ModelType.FT,
|
| 396 |
+
"Neko-Institute-of-Science/pygmalion-7b": ModelType.FT,
|
| 397 |
+
"SebastianSchramm/Cerebras-GPT-111M-instruction": ModelType.FT,
|
| 398 |
+
"victor123/WizardLM-13B-1.0": ModelType.FT,
|
| 399 |
+
"OpenBuddy/openbuddy-openllama-13b-v7-fp16": ModelType.FT,
|
| 400 |
+
"OpenBuddy/openbuddy-llama2-13b-v8.1-fp16": ModelType.FT,
|
| 401 |
+
"OpenBuddyEA/openbuddy-llama-30b-v7.1-bf16": ModelType.FT,
|
| 402 |
"baichuan-inc/Baichuan-7B": ModelType.PT,
|
| 403 |
+
"tiiuae/falcon-40b-instruct": ModelType.FT,
|
| 404 |
"tiiuae/falcon-40b": ModelType.PT,
|
| 405 |
"tiiuae/falcon-7b": ModelType.PT,
|
| 406 |
+
"YeungNLP/firefly-llama-13b": ModelType.FT,
|
| 407 |
+
"YeungNLP/firefly-llama-13b-v1.2": ModelType.FT,
|
| 408 |
+
"YeungNLP/firefly-llama2-13b": ModelType.FT,
|
| 409 |
+
"YeungNLP/firefly-ziya-13b": ModelType.FT,
|
| 410 |
+
"shaohang/Sparse0.5_OPT-1.3": ModelType.FT,
|
| 411 |
+
"xzuyn/Alpacino-SuperCOT-13B": ModelType.FT,
|
| 412 |
+
"xzuyn/MedicWizard-7B": ModelType.FT,
|
| 413 |
+
"xDAN-AI/xDAN_13b_l2_lora": ModelType.FT,
|
| 414 |
+
"beomi/KoAlpaca-Polyglot-5.8B": ModelType.FT,
|
| 415 |
+
"beomi/llama-2-ko-7b": ModelType.FT,
|
| 416 |
"Salesforce/codegen-6B-multi": ModelType.PT,
|
| 417 |
"Salesforce/codegen-16B-nl": ModelType.PT,
|
| 418 |
"Salesforce/codegen-6B-nl": ModelType.PT,
|
| 419 |
+
"ai-forever/rugpt3large_based_on_gpt2": ModelType.FT,
|
| 420 |
"gpt2-large": ModelType.PT,
|
| 421 |
+
"frank098/orca_mini_3b_juniper": ModelType.FT,
|
| 422 |
+
"frank098/WizardLM_13B_juniper": ModelType.FT,
|
| 423 |
+
"FPHam/Free_Sydney_13b_HF": ModelType.FT,
|
| 424 |
"huggingface/llama-13b": ModelType.PT,
|
| 425 |
"huggingface/llama-7b": ModelType.PT,
|
| 426 |
"huggingface/llama-65b": ModelType.PT,
|
| 427 |
"huggingface/llama-65b": ModelType.PT,
|
| 428 |
"huggingface/llama-30b": ModelType.PT,
|
| 429 |
+
"Henk717/chronoboros-33B": ModelType.FT,
|
| 430 |
+
"jondurbin/airoboros-13b-gpt4-1.4": ModelType.FT,
|
| 431 |
+
"jondurbin/airoboros-7b": ModelType.FT,
|
| 432 |
+
"jondurbin/airoboros-7b-gpt4": ModelType.FT,
|
| 433 |
+
"jondurbin/airoboros-7b-gpt4-1.1": ModelType.FT,
|
| 434 |
+
"jondurbin/airoboros-7b-gpt4-1.2": ModelType.FT,
|
| 435 |
+
"jondurbin/airoboros-7b-gpt4-1.3": ModelType.FT,
|
| 436 |
+
"jondurbin/airoboros-7b-gpt4-1.4": ModelType.FT,
|
| 437 |
+
"jondurbin/airoboros-l2-7b-gpt4-1.4.1": ModelType.FT,
|
| 438 |
+
"jondurbin/airoboros-l2-13b-gpt4-1.4.1": ModelType.FT,
|
| 439 |
+
"jondurbin/airoboros-l2-70b-gpt4-1.4.1": ModelType.FT,
|
| 440 |
+
"jondurbin/airoboros-13b": ModelType.FT,
|
| 441 |
+
"jondurbin/airoboros-33b-gpt4-1.4": ModelType.FT,
|
| 442 |
+
"jondurbin/airoboros-33b-gpt4-1.2": ModelType.FT,
|
| 443 |
+
"jondurbin/airoboros-65b-gpt4-1.2": ModelType.FT,
|
| 444 |
+
"ariellee/SuperPlatty-30B": ModelType.FT,
|
| 445 |
+
"danielhanchen/open_llama_3b_600bt_preview": ModelType.FT,
|
| 446 |
"cerebras/Cerebras-GPT-256M": ModelType.PT,
|
| 447 |
"cerebras/Cerebras-GPT-1.3B": ModelType.PT,
|
| 448 |
"cerebras/Cerebras-GPT-13B": ModelType.PT,
|
|
|
|
| 450 |
"cerebras/Cerebras-GPT-111M": ModelType.PT,
|
| 451 |
"cerebras/Cerebras-GPT-6.7B": ModelType.PT,
|
| 452 |
"Yhyu13/oasst-rlhf-2-llama-30b-7k-steps-hf": ModelType.RL,
|
| 453 |
+
"Yhyu13/llama-30B-hf-openassitant": ModelType.FT,
|
| 454 |
+
"NousResearch/Nous-Hermes-Llama2-13b": ModelType.FT,
|
| 455 |
+
"NousResearch/Nous-Hermes-llama-2-7b": ModelType.FT,
|
| 456 |
+
"NousResearch/Redmond-Puffin-13B": ModelType.FT,
|
| 457 |
+
"NousResearch/Nous-Hermes-13b": ModelType.FT,
|
| 458 |
+
"project-baize/baize-v2-7b": ModelType.FT,
|
| 459 |
+
"project-baize/baize-v2-13b": ModelType.FT,
|
| 460 |
+
"LLMs/WizardLM-13B-V1.0": ModelType.FT,
|
| 461 |
+
"LLMs/AlpacaGPT4-7B-elina": ModelType.FT,
|
| 462 |
+
"wenge-research/yayi-7b": ModelType.FT,
|
| 463 |
+
"wenge-research/yayi-7b-llama2": ModelType.FT,
|
| 464 |
+
"wenge-research/yayi-13b-llama2": ModelType.FT,
|
| 465 |
+
"yhyhy3/open_llama_7b_v2_med_instruct": ModelType.FT,
|
| 466 |
+
"llama-anon/instruct-13b": ModelType.FT,
|
| 467 |
+
"huggingtweets/jerma985": ModelType.FT,
|
| 468 |
+
"huggingtweets/gladosystem": ModelType.FT,
|
| 469 |
+
"huggingtweets/bladeecity-jerma985": ModelType.FT,
|
| 470 |
"huggyllama/llama-13b": ModelType.PT,
|
| 471 |
"huggyllama/llama-65b": ModelType.PT,
|
| 472 |
"FabbriSimo01/Facebook_opt_1.3b_Quantized": ModelType.PT,
|
| 473 |
+
"upstage/Llama-2-70b-instruct": ModelType.FT,
|
| 474 |
+
"upstage/Llama-2-70b-instruct-1024": ModelType.FT,
|
| 475 |
+
"upstage/llama-65b-instruct": ModelType.FT,
|
| 476 |
+
"upstage/llama-30b-instruct-2048": ModelType.FT,
|
| 477 |
+
"upstage/llama-30b-instruct": ModelType.FT,
|
| 478 |
+
"WizardLM/WizardLM-13B-1.0": ModelType.FT,
|
| 479 |
+
"WizardLM/WizardLM-13B-V1.1": ModelType.FT,
|
| 480 |
+
"WizardLM/WizardLM-13B-V1.2": ModelType.FT,
|
| 481 |
+
"WizardLM/WizardLM-30B-V1.0": ModelType.FT,
|
| 482 |
+
"WizardLM/WizardCoder-15B-V1.0": ModelType.FT,
|
| 483 |
"gpt2": ModelType.PT,
|
| 484 |
+
"keyfan/vicuna-chinese-replication-v1.1": ModelType.FT,
|
| 485 |
+
"nthngdy/pythia-owt2-70m-100k": ModelType.FT,
|
| 486 |
+
"nthngdy/pythia-owt2-70m-50k": ModelType.FT,
|
| 487 |
+
"quantumaikr/KoreanLM-hf": ModelType.FT,
|
| 488 |
+
"quantumaikr/open_llama_7b_hf": ModelType.FT,
|
| 489 |
+
"quantumaikr/QuantumLM-70B-hf": ModelType.FT,
|
| 490 |
+
"MayaPH/FinOPT-Lincoln": ModelType.FT,
|
| 491 |
+
"MayaPH/FinOPT-Franklin": ModelType.FT,
|
| 492 |
+
"MayaPH/GodziLLa-30B": ModelType.FT,
|
| 493 |
+
"MayaPH/GodziLLa-30B-plus": ModelType.FT,
|
| 494 |
+
"MayaPH/FinOPT-Washington": ModelType.FT,
|
| 495 |
+
"ogimgio/gpt-neo-125m-neurallinguisticpioneers": ModelType.FT,
|
| 496 |
+
"layoric/llama-2-13b-code-alpaca": ModelType.FT,
|
| 497 |
+
"CobraMamba/mamba-gpt-3b": ModelType.FT,
|
| 498 |
+
"CobraMamba/mamba-gpt-3b-v2": ModelType.FT,
|
| 499 |
+
"CobraMamba/mamba-gpt-3b-v3": ModelType.FT,
|
| 500 |
+
"timdettmers/guanaco-33b-merged": ModelType.FT,
|
| 501 |
+
"elinas/chronos-33b": ModelType.FT,
|
| 502 |
+
"heegyu/RedTulu-Uncensored-3B-0719": ModelType.FT,
|
| 503 |
+
"heegyu/WizardVicuna-Uncensored-3B-0719": ModelType.FT,
|
| 504 |
+
"heegyu/WizardVicuna-3B-0719": ModelType.FT,
|
| 505 |
"meta-llama/Llama-2-7b-chat-hf": ModelType.RL,
|
| 506 |
"meta-llama/Llama-2-7b-hf": ModelType.PT,
|
| 507 |
"meta-llama/Llama-2-13b-chat-hf": ModelType.RL,
|
| 508 |
"meta-llama/Llama-2-13b-hf": ModelType.PT,
|
| 509 |
"meta-llama/Llama-2-70b-chat-hf": ModelType.RL,
|
| 510 |
"meta-llama/Llama-2-70b-hf": ModelType.PT,
|
| 511 |
+
"xhyi/PT_GPTNEO350_ATG": ModelType.FT,
|
| 512 |
+
"h2oai/h2ogpt-gm-oasst1-en-1024-20b": ModelType.FT,
|
| 513 |
+
"h2oai/h2ogpt-gm-oasst1-en-1024-open-llama-7b-preview-400bt": ModelType.FT,
|
| 514 |
+
"h2oai/h2ogpt-oig-oasst1-512-6_9b": ModelType.FT,
|
| 515 |
+
"h2oai/h2ogpt-oasst1-512-12b": ModelType.FT,
|
| 516 |
+
"h2oai/h2ogpt-oig-oasst1-256-6_9b": ModelType.FT,
|
| 517 |
+
"h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt": ModelType.FT,
|
| 518 |
+
"h2oai/h2ogpt-oasst1-512-20b": ModelType.FT,
|
| 519 |
+
"h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt-v2": ModelType.FT,
|
| 520 |
+
"h2oai/h2ogpt-gm-oasst1-en-1024-12b": ModelType.FT,
|
| 521 |
+
"h2oai/h2ogpt-gm-oasst1-multilang-1024-20b": ModelType.FT,
|
| 522 |
+
"bofenghuang/vigogne-13b-instruct": ModelType.FT,
|
| 523 |
+
"bofenghuang/vigogne-13b-chat": ModelType.FT,
|
| 524 |
+
"bofenghuang/vigogne-2-7b-instruct": ModelType.FT,
|
| 525 |
+
"bofenghuang/vigogne-7b-instruct": ModelType.FT,
|
| 526 |
+
"bofenghuang/vigogne-7b-chat": ModelType.FT,
|
| 527 |
+
"Vmware/open-llama-7b-v2-open-instruct": ModelType.FT,
|
| 528 |
+
"VMware/open-llama-0.7T-7B-open-instruct-v1.1": ModelType.FT,
|
| 529 |
+
"ewof/koishi-instruct-3b": ModelType.FT,
|
| 530 |
+
"gywy/llama2-13b-chinese-v1": ModelType.FT,
|
| 531 |
+
"GOAT-AI/GOAT-7B-Community": ModelType.FT,
|
| 532 |
+
"psyche/kollama2-7b": ModelType.FT,
|
| 533 |
+
"TheTravellingEngineer/llama2-7b-hf-guanaco": ModelType.FT,
|
| 534 |
+
"beaugogh/pythia-1.4b-deduped-sharegpt": ModelType.FT,
|
| 535 |
+
"augtoma/qCammel-70-x": ModelType.FT,
|
| 536 |
+
"Lajonbot/Llama-2-7b-chat-hf-instruct-pl-lora_unload": ModelType.FT,
|
| 537 |
+
"anhnv125/pygmalion-6b-roleplay": ModelType.FT,
|
| 538 |
+
"64bits/LexPodLM-13B": ModelType.FT
|
| 539 |
}
|
| 540 |
|
| 541 |
|
|
|
|
| 551 |
elif model_data[AutoEvalColumn.model_type.name] == "pretrained" or any([i in model_data["model_name_for_query"] for i in ["pretrained"]]):
|
| 552 |
model_data[AutoEvalColumn.model_type.name] = ModelType.PT.value.name
|
| 553 |
model_data[AutoEvalColumn.model_type_symbol.name] = ModelType.PT.value.symbol + ("🔺" if is_delta else "")
|
| 554 |
+
elif model_data[AutoEvalColumn.model_type.name] == "instruct" or any([i in model_data["model_name_for_query"] for i in ["instruct", "-ift-"]]):
|
| 555 |
+
model_data[AutoEvalColumn.model_type.name] = ModelType.IFT.value.name
|
| 556 |
+
model_data[AutoEvalColumn.model_type_symbol.name] = ModelType.IFT.value.symbol + ("🔺" if is_delta else "")
|
| 557 |
elif model_data[AutoEvalColumn.model_type.name] == "finetuned" or any([i in model_data["model_name_for_query"] for i in ["finetuned", "-ft-"]]):
|
| 558 |
+
model_data[AutoEvalColumn.model_type.name] = ModelType.FT.value.name
|
| 559 |
+
model_data[AutoEvalColumn.model_type_symbol.name] = ModelType.FT.value.symbol + ("🔺" if is_delta else "")
|
| 560 |
elif model_data[AutoEvalColumn.model_type.name] == "with RL" or any([i in model_data["model_name_for_query"] for i in ["-rl-", "-rlhf-"]]):
|
| 561 |
model_data[AutoEvalColumn.model_type.name] = ModelType.RL.value.name
|
| 562 |
model_data[AutoEvalColumn.model_type_symbol.name] = ModelType.RL.value.symbol + ("🔺" if is_delta else "")
|