|
--- |
|
base_model: HuggingFaceH4/zephyr-7b-beta |
|
inference: false |
|
license: mit |
|
model_creator: HuggingFaceH4 |
|
model_name: Zephyr 7B Beta |
|
model_type: mistral |
|
prompt_template: "<|system|> |
|
</s> |
|
<|user|> |
|
{prompt}</s> |
|
<|assistant|> |
|
" |
|
quantized_by: Semantically AI |
|
pruned_by: Semantically AI |
|
--- |
|
|
|
# Zephyr 7B Beta - GGUF |
|
|
|
- Model creator: [Hugging Face H4](https://huggingface.co/HuggingFaceH4) |
|
- Original model: [Zephyr 7B Beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) |
|
|
|
<!-- description start --> |
|
|
|
## Description |
|
|
|
This repo contains GGUF format model files for [Hugging Face H4's Zephyr 7B Beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta). |
|
|
|
<!-- description end --> |
|
<!-- README_GGUF.md-about-gguf start --> |
|
|
|
GGUF is a new format introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp. |
|
|
|
<!-- README_GGUF.md-about-gguf end --> |
|
<!-- prompt-template start --> |
|
|
|
## Prompt template: Zephyr |
|
|
|
``` |
|
<|system|> |
|
</s> |
|
<|user|> |
|
{prompt}</s> |
|
<|assistant|> |
|
``` |
|
|
|
<!-- prompt-template end --> |
|
|
|
<!-- explaination start --> |
|
|
|
## Explanation of quantisation methods |
|
|
|
<details> |
|
<summary>Click to see details</summary> |
|
|
|
The GGUF model is pruned to 50% using sparseGPT method [sparseGPT](https://github.com/IST-DASLab/sparsegpt) |
|
|
|
</details> |
|
<!-- explaination end --> |
|
|
|
``` |
|
from llama_cpp import Llama |
|
|
|
llm = Llama(model_path="zephyr-7b-beta-pruned50-Q8_0.gguf") |
|
|
|
output = llm(""" |
|
<|system|> |
|
You are a friendly chatbot who always responds in the style of a pirate.</s> |
|
<|user|> |
|
How many helicopters can a human eat in one sitting?</s> |
|
<|assistant|>""") |
|
|
|
print(output) |
|
# |
|
``` |
|
|
|
<!-- README_GGUF.md-how-to-run start --> |
|
|