--- library_name: transformers tags: [] --- # Model Card for Model ID This is a testing model by: - **shrinking the layers of llama-3-8b to only 2 layers of transformer decoder** - **adding a customized layer to llama attention** - **shrinking the total param size to around 2b... so pathetic** 😢😢😢 The purpose of this model is to show how you can download a pre-trained llama and customize it... however you want... and then re-train it with... whatever you want... and then upload your model to hugging face 🤗 It is not intended to compete with larger models developed by large corporations with GPU advantage. However, of course, if you want to use this model properly, you have to not only download it with LlamaForCausalLM but also to use the following code in .ipynb to build the same customized layer and merge the huggingface safetensor to each parameter: - **Repository:** https://github.com/Brownwang0426/Cullama # Model Details - **Developed by:** Po-Lung Wang - **Model type:** Llama-3 - **License:** MIT