|
--- |
|
base_model: jpacifico/Chocolatine-3B-Instruct-DPO-v1.2 |
|
datasets: |
|
- jpacifico/french-orca-dpo-pairs-revised |
|
language: |
|
- fr |
|
- en |
|
library_name: transformers |
|
license: mit |
|
pipeline_tag: text-generation |
|
tags: |
|
- french |
|
- chocolatine |
|
- llama-cpp |
|
--- |
|
|
|
# Chocolatine-3B-Instruct-DPO-v1.2-Q4_K_M-GGUF |
|
|
|
Quantized q4_k_m GGUF version of the original model [`jpacifico/Chocolatine-3B-Instruct-DPO-v1.2`](https://huggingface.co/jpacifico/Chocolatine-3B-Instruct-DPO-v1.2) |
|
can be used on a CPU device, compatible [llama.cpp](https://github.com/ggerganov/llama.cpp) |
|
now supported architecture by [LM Studio](https://lmstudio.ai/). |
|
Also ready for Raspberry Pi 5 8Gb. |
|
|
|
### Ollama |
|
|
|
[jpacifico/chocolatine-3b](https://ollama.com/jpacifico/chocolatine-3b) |
|
|
|
Usage: |
|
|
|
```bash |
|
ollama run jpacifico/chocolatine-3b |
|
``` |
|
|
|
Ollama *Modelfile* example : |
|
|
|
```bash |
|
FROM ./chocolatine-3b-instruct-dpo-v1.2-q4_k_m.gguf |
|
TEMPLATE """{{ if .System }}<|system|> |
|
{{ .System }}<|end|> |
|
{{ end }}{{ if .Prompt }}<|user|> |
|
{{ .Prompt }}<|end|> |
|
{{ end }}<|assistant|> |
|
{{ .Response }}<|end|> |
|
""" |
|
PARAMETER stop """{"stop": ["<|end|>","<|user|>","<|assistant|>"]}""" |
|
SYSTEM """You are a friendly assistant called Chocolatine.""" |
|
``` |
|
|
|
|
|
### Limitations |
|
|
|
The Chocolatine model is a quick demonstration that a base model can be easily fine-tuned to achieve compelling performance. |
|
It does not have any moderation mechanism. |
|
|
|
- **Developed by:** Jonathan Pacifico, 2024 |
|
- **Model type:** LLM |
|
- **Language(s) (NLP):** French, English |
|
- **License:** MIT |
|
|