RedHatAI/Mixtral-8x7B-Instruct-v0.1-AutoFP8
Text Generation
•
47B
•
Updated
•
23
•
3
RedHatAI/Meta-Llama-3-70B-Instruct-FP8
Text Generation
•
71B
•
Updated
•
58.6k
•
13
RedHatAI/Meta-Llama-3-8B-Instruct-FP8
Text Generation
•
8B
•
Updated
•
2.94k
•
23
RedHatAI/DeepSeek-Coder-V2-Lite-Base-FP8
Text Generation
•
16B
•
Updated
•
52
RedHatAI/DeepSeek-Coder-V2-Lite-Instruct-FP8
Text Generation
•
16B
•
Updated
•
23.6k
•
7
RedHatAI/Qwen2-7B-Instruct-quantized.w4a16
Text Generation
•
2B
•
Updated
•
46
RedHatAI/Qwen2-72B-Instruct-quantized.w4a16
Text Generation
•
12B
•
Updated
•
12
•
4
RedHatAI/Qwen2-1.5B-Instruct-quantized.w4a16
Text Generation
•
0.6B
•
Updated
•
24
RedHatAI/Qwen2-0.5B-Instruct-quantized.w4a16
Text Generation
•
0.3B
•
Updated
•
16
RedHatAI/Qwen2-72B-Instruct-quantized.w8a16
Text Generation
•
20B
•
Updated
•
680
•
1
RedHatAI/Qwen2-7B-Instruct-quantized.w8a16
Text Generation
•
3B
•
Updated
•
31
RedHatAI/Qwen2-1.5B-Instruct-quantized.w8a16
Text Generation
•
0.6B
•
Updated
•
12
RedHatAI/Qwen2-0.5B-Instruct-quantized.w8a16
Text Generation
•
0.2B
•
Updated
•
12
RedHatAI/Llama-2-7b-chat-quantized.w4a16
Text Generation
•
1B
•
Updated
•
16
RedHatAI/Meta-Llama-3-8B-Instruct-quantized.w4a16
Text Generation
•
2B
•
Updated
•
125
•
2
RedHatAI/Llama-2-7b-chat-quantized.w8a16
Text Generation
•
2B
•
Updated
•
15
RedHatAI/Mistral-7B-Instruct-v0.3-quantized.w8a16
Text Generation
•
2B
•
Updated
•
273
RedHatAI/Meta-Llama-3-70B-Instruct-quantized.w8a16
Text Generation
•
19B
•
Updated
•
1.12k
•
5
RedHatAI/Meta-Llama-3-8B-Instruct-quantized.w8a16
Text Generation
•
3B
•
Updated
•
1.34k
•
3
RedHatAI/SparseLLama-2-7b-ultrachat_200k-pruned_50.2of4
Text Generation
•
7B
•
Updated
•
13
RedHatAI/SparseLlama-2-7b-evolcodealpaca-pruned_50.2of4
Text Generation
•
7B
•
Updated
•
11
RedHatAI/Meta-Llama-3-70B-Instruct-FP8-KV
Text Generation
•
71B
•
Updated
•
52
•
2
RedHatAI/Llama-2-7b-gsm8k-pruned_70
Text Generation
•
7B
•
Updated
•
16
RedHatAI/Llama-2-7b-gsm8k-pruned_50
Text Generation
•
7B
•
Updated
•
16
•
1
RedHatAI/Llama-2-7b-gsm8k
Text Generation
•
Updated
•
1.71k
•
3
RedHatAI/Meta-Llama-3-8B-Instruct-FP8-KV
Text Generation
•
8B
•
Updated
•
4.56k
•
8
RedHatAI/Mistral-7B-Instruct-v0.3-GPTQ-4bit
Text Generation
•
1B
•
Updated
•
8.72k
•
19
RedHatAI/SparseLlama-2-7b-cnn-daily-mail-pruned_50.2of4
Text Generation
•
7B
•
Updated
•
11
RedHatAI/SparseLlama-2-7b-cnn-daily-mail-pruned_70
Updated
RedHatAI/Llama-2-7b-cnn-daily-mail-pruned_70-quantized-deepsparse
Text Generation
•
Updated
•
12