This is a quick GGUF quantization of DiscoResearch/Llama3-DiscoLeo-Instruct-8B-32k-v0.1 (done for testing purposes with an older llama.cpp version without bpe pre-tokenizer fix)

Downloads last month
1
GGUF
Model size
8.03B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support