Edit model card

Original Model: https://huggingface.co/aetherwiing/MN-12B-Starcannon-v2

made with https://huggingface.co/FantasiaFoundry/GGUF-Quantization-Script

Models Q2_K_L, Q4_K_L, Q5_K_L, Q6_K_L, are using Q_8 output tensors and token embeddings

using bartowski's imatrix dataset

(happened to already have made these before mradermacher, probably use theirs instead)

Downloads last month
13
GGUF
Model size
12.2B params
Architecture
llama

4-bit

6-bit

Inference Examples
Inference API (serverless) is not available, repository is disabled.

Model tree for Reiterate3680/MN-12B-Starcannon-v2-GGUF

Quantized
this model