Edit model card

Original model: https://huggingface.co/google/gemma-2-2b-jpn-it

Run them in LM Studio

Prompt format

<|system|> {system_prompt}<|end|><|user|> {prompt}<|end|><|assistant|>

Download a file (not the whole branch) from below:

Filename Quant type File Size Split Description
gemma-2-2b-jpn-it-f16.gguf f16 5.24GB false Full F16 weights.
gemma-2-2b-jpn-it-Q8_0.gguf Q8_0 2.78GB false Extremely high quality, recommended.

Downloading using huggingface-cli

First, make sure you have hugginface-cli installed:

pip install -U "huggingface_hub[cli]"

Then, you can target the specific file you want:

huggingface-cli download ymcki/gemma-2-2b-jpn-it-GGUF --include "gemma-2-2b-jpn-it-Q8_0.gguf" --local-dir ./

Credits

Thank you bartowski for providing a README.md to get me started.

Downloads last month
26
GGUF
Model size
2.61B params
Architecture
gemma2

8-bit

16-bit

Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for ymcki/gemma-2-2b-jpn-it-GGUF

Base model

google/gemma-2-2b
Quantized
(10)
this model