license: gpl-3.0 | |
A Chinese instruction-tuned LLaMA(30b) | |
dataset: translated alpaca instruction dataset. | |
# Usage | |
Please check [Alpaca](https://github.com/tloen/alpaca-lora) to install the base project and then | |
``` | |
python generate.py \ | |
--load_8bit \ | |
--base_model 'decapoda-research/llama-30b-hf' \ | |
--lora_weights 'llmatics/alpaca-lora-cn-30b' | |
``` |