This is a quick GGUF quantization of DiscoResearch/Llama3-DiscoLeo-Instruct-8B-32k-v0.1 (done for testing purposes with an older llama.cpp version without bpe pre-tokenizer fix)
- Downloads last month
- 1
Hardware compatibility
Log In
to view the estimation
4-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support