File size: 7,174 Bytes
bb98a44
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
MODEL_MAP = {
  "mixtral-8x7b": "mistralai/Mixtral-8x7B-Instruct-v0.1",
  "default": "meta-llama/Llama-3.3-70B-Instruct",
  "Llama-3.3-70B-Instruct": "meta-llama/Llama-3.3-70B-Instruct",
  "Qwen2.5-72B-Instruct": "Qwen/Qwen2.5-72B-Instruct",
  "c4ai-command-r-plus-08-2024": "CohereForAI/c4ai-command-r-plus-08-2024",
  "DeepSeek-R1-Distill-Qwen-32B": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B",
  "Llama-3.1-Nemotron-70B-Instruct-HF": "nvidia/Llama-3.1-Nemotron-70B-Instruct-HF",
  "QwQ-32B-Preview": "Qwen/QwQ-32B-Preview",
  "Qwen2.5-Coder-32B-Instruct": "Qwen/Qwen2.5-Coder-32B-Instruct",
  "Llama-3.2-11B-Vision-Instruct": "meta-llama/Llama-3.2-11B-Vision-Instruct",
  "Hermes-3-Llama-3.1-8B": "NousResearch/Hermes-3-Llama-3.1-8B",
  "Mistral-Nemo-Instruct-2407": "mistralai/Mistral-Nemo-Instruct-2407",
  "Phi-3.5-mini-instruct": "microsoft/Phi-3.5-mini-instruct",
    "nous-mixtral-8x7b": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
    "mistral-7b": "mistralai/Mistral-7B-Instruct-v0.2",
    "yi-1.5-34b": "01-ai/Yi-1.5-34B-Chat",
    "gemma-7b": "google/gemma-1.1-7b-it",
}
# {
#     "mixtral-8x7b": "mistralai/Mixtral-8x7B-Instruct-v0.1",  # [Recommended]
#     "nous-mixtral-8x7b": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
#     "mistral-7b": "mistralai/Mistral-7B-Instruct-v0.2",
#     "yi-1.5-34b": "01-ai/Yi-1.5-34B-Chat",
#     "gemma-7b": "google/gemma-1.1-7b-it",
#     # "openchat-3.5": "openchat/openchat-3.5-0106",
#     # "command-r-plus": "CohereForAI/c4ai-command-r-plus",
#     # "llama3-70b": "meta-llama/Meta-Llama-3-70B-Instruct",
#     # "zephyr-141b": "HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1",
#     "default": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
# }

AVAILABLE_MODELS = list(MODEL_MAP.keys())

PRO_MODELS = ["command-r-plus", "llama3-70b", "zephyr-141b"]

STOP_SEQUENCES_MAP = {
    # https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1/blob/main/tokenizer_config.json#L33
    "mixtral-8x7b": "</s>",
    # https://huggingface.co/NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO/blob/main/tokenizer_config.json#L50
    "nous-mixtral-8x7b": "<|im_end|>",
    # https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2/blob/main/tokenizer_config.json#L33
    "mistral-7b": "</s>",
    # https://huggingface.co/01-ai/Yi-1.5-34B-Chat/blob/main/tokenizer_config.json#L42
    "yi-1.5-34b": "<|im_end|>",
    # https://huggingface.co/google/gemma-1.1-7b-it/blob/main/tokenizer_config.json#L1509
    "gemma-7b": "<eos>",
    # https://huggingface.co/huihui-ai/Llama-3.3-70B-Instruct-abliterated/blob/fa13334669544bab573e0e5313cad629a9c02e2c/tokenizer_config.json#L2055
    "Llama-3.3-70B-Instruct": "<|eot_id|>",
    # https://huggingface.co/google/gemma-1.1-7b-it/blob/main/tokenizer_config.json#L1509
    "c4ai-command-r-plus-08-2024": "<|END_OF_TURN_TOKEN|>",
    # https://huggingface.co/unsloth/Qwen2.5-72B-Instruct/blob/main/tokenizer_config.json
    "Qwen2.5-72B-Instruct":"<|im_end|>",
    # https://huggingface.co/unsloth/DeepSeek-R1-Distill-Qwen-32B/blob/main/tokenizer_config.json
    "DeepSeek-R1-Distill-Qwen-32B":"<|end▁of▁sentence|>",

    # "openchat-3.5": "<|end_of_turn|>",
    # "command-r-plus": "<|END_OF_TURN_TOKEN|>",
}

TOKEN_LIMIT_MAP = {
    "mixtral-8x7b": 32768,
    "nous-mixtral-8x7b": 32768,
    "mistral-7b": 32768,
    "yi-1.5-34b": 4096,
    "gemma-7b": 8192,
    "Llama-3.3-70B-Instruct": 131072,
    "c4ai-command-r-plus-08-2024":1000000000000000019884624838656,
    "Qwen2.5-72B-Instruct":131072,
    "DeepSeek-R1-Distill-Qwen-32B":131072,

    # "openchat-3.5": 8192,
    # "command-r-plus": 32768,
    # "llama3-70b": 8192,
    # "zephyr-141b": 2048,
    # "gpt-3.5-turbo": 8192,
}

TOKEN_RESERVED = 20


# https://platform.openai.com/docs/api-reference/models/list
AVAILABLE_MODELS_DICTS = [
  {
    "id": "Llama-3.3-70B-Instruct",
    "description": "[meta-llama/Llama-3.3-70B-Instruct]: https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct",
    "object": "model",
    "created": 1700000000,
    "owned_by": "meta-llama"
  },
  {
    "id": "Qwen2.5-72B-Instruct",
    "description": "[Qwen/Qwen2.5-72B-Instruct]: https://huggingface.co/Qwen/Qwen2.5-72B-Instruct",
    "object": "model",
    "created": 1700000000,
    "owned_by": "Qwen"
  },
  {
    "id": "c4ai-command-r-plus-08-2024",
    "description": "[CohereForAI/c4ai-command-r-plus-08-2024]: https://huggingface.co/CohereForAI/c4ai-command-r-plus-08-2024",
    "object": "model",
    "created": 1700000000,
    "owned_by": "CohereForAI"
  },
  {
    "id": "DeepSeek-R1-Distill-Qwen-32B",
    "description": "[deepseek-ai/DeepSeek-R1-Distill-Qwen-32B]: https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B",
    "object": "model",
    "created": 1700000000,
    "owned_by": "deepseek-ai"
  },
  {
    "id": "Llama-3.1-Nemotron-70B-Instruct-HF",
    "description": "[nvidia/Llama-3.1-Nemotron-70B-Instruct-HF]: https://huggingface.co/nvidia/Llama-3.1-Nemotron-70B-Instruct-HF",
    "object": "model",
    "created": 1700000000,
    "owned_by": "nvidia"
  },
  {
    "id": "QwQ-32B-Preview",
    "description": "[Qwen/QwQ-32B-Preview]: https://huggingface.co/Qwen/QwQ-32B-Preview",
    "object": "model",
    "created": 1700000000,
    "owned_by": "Qwen"
  },
  {
    "id": "Qwen2.5-Coder-32B-Instruct",
    "description": "[Qwen/Qwen2.5-Coder-32B-Instruct]: https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct",
    "object": "model",
    "created": 1700000000,
    "owned_by": "Qwen"
  },
  {
    "id": "Llama-3.2-11B-Vision-Instruct",
    "description": "[meta-llama/Llama-3.2-11B-Vision-Instruct]: https://huggingface.co/meta-llama/Llama-3.2-11B-Vision-Instruct",
    "object": "model",
    "created": 1700000000,
    "owned_by": "meta-llama"
  },
  {
    "id": "Hermes-3-Llama-3.1-8B",
    "description": "[NousResearch/Hermes-3-Llama-3.1-8B]: https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-8B",
    "object": "model",
    "created": 1700000000,
    "owned_by": "NousResearch"
  },
  {
    "id": "Mistral-Nemo-Instruct-2407",
    "description": "[mistralai/Mistral-Nemo-Instruct-2407]: https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407",
    "object": "model",
    "created": 1700000000,
    "owned_by": "mistralai"
  },
  {
    "id": "Phi-3.5-mini-instruct",
    "description": "[microsoft/Phi-3.5-mini-instruct]: https://huggingface.co/microsoft/Phi-3.5-mini-instruct",
    "object": "model",
    "created": 1700000000,
    "owned_by": "microsoft"
  }
]


'''
https://huggingface.co/chat/models
var links = document.querySelectorAll('div.relative a');
var mlist=[];
links.forEach(link=>{
  mlist.push(link.href.split('https://huggingface.co/chat/models/')[1])
})

p_MODEL_MAP(mlist);
p_AVAILABLE_MODELS_DICTS(mlist);


function p_MODEL_MAP(mlist){
  let d = {}
  mlist.forEach(m=>{
    d[m.split('/')[1]]=m
  });
  console.log(d)
  return d
}



function p_AVAILABLE_MODELS_DICTS(mlist){
  let d = []
  mlist.forEach(m=>{
    d.push({
        "id": m.split('/')[1],
        "description": `[${m}]: https://huggingface.co/${m}`,
        "object": "model",
        "created": 1700000000,
        "owned_by": m.split('/')[0],
    },)
     
  });
  console.log(d)
  return d
}

'''