QuantFactory Banner

QuantFactory/MN-12B-Vespa-x1-GGUF

This is quantized version of Sao10K/MN-12B-Vespa-x1 created using llama.cpp

Original Model Card

Mistral-NeMo-12B-Vespa-x1

Vespa


This is an experiment.

Like Hanami, both are not made to be a roleplaying model, but they can roleplay.

Format, chatml?

Temp <1, min_p of 0.1 atleast. ty.

I like it, so try it out?

Downloads last month
17
GGUF
Model size
9.24B params
Architecture
gemma2
Hardware compatibility
Log In to view the estimation

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support