QuantFactory/MN-12B-Vespa-x1-GGUF
This is quantized version of Sao10K/MN-12B-Vespa-x1 created using llama.cpp
Original Model Card
Mistral-NeMo-12B-Vespa-x1
This is an experiment.
Like Hanami, both are not made to be a roleplaying model, but they can roleplay.
Format, chatml?
Temp <1, min_p of 0.1 atleast. ty.
I like it, so try it out?
- Downloads last month
- 17
Hardware compatibility
Log In
to view the estimation
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support