aminabbasi commited on
Commit
1b5ae91
1 Parent(s): c3cab1c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +61 -3
README.md CHANGED
@@ -1,3 +1,61 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ language:
4
+ - fa
5
+ library_name: transformers
6
+ tags:
7
+ - text-generation-inference
8
+ inference: false
9
+ pipeline_tag: text-generation
10
+ ---
11
+
12
+ # PersianLLaMA: Towards Building First Persian Large Language Model
13
+
14
+ ## 🌟 Introduction
15
+ Welcome to the home of PersianLLaMA, the pioneering large language model for the Persian language. With 13 billion parameters, this model is trained on a diverse corpus and designed to excel in multiple NLP tasks, setting a new benchmark for Persian language understanding and generation.
16
+
17
+ ![PersianLLaMA](https://huggingface.co/ViraIntelligentDataMining/PersianLLaMA-2-13B/resolve/main/persianllama.png)
18
+
19
+ ## 🛠 Model Description
20
+ PersianLLaMA is not just a model but a comprehensive tool for:
21
+ - 📝 **Text Generation**: Crafting coherent and contextually appropriate text.
22
+ - 🎯 **Instruct Tuning**: Executing tasks based on detailed instructions, ideal for scenarios where the model needs to adhere to specific guidelines or produce outputs tailored to particular requirements.
23
+ - ❓ **Question Answering**: Providing accurate answers to Persian queries.
24
+ - 📊 **Text Summarization**: Condensing Persian texts into precise summaries.
25
+
26
+ This model has been collaboratively developed by a team of experts, including Mohammad Amin Abbasi, Arash Ghafouri, Mahdi Firouzmandi, Hassan Naderi, and Behrouz Minaei Bidgoli.
27
+
28
+ ## 🚀 Quick Start
29
+ To integrate PersianLLaMA into your project, follow these steps:
30
+ ```python
31
+ from transformers import AutoModelForCausalLM, AutoTokenizer
32
+
33
+ model_name = "your-huggingface-username/persianllama"
34
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
35
+ model = AutoModelForCausalLM.from_pretrained(model_name)
36
+
37
+ prompt = "این متن به فارسی است"
38
+ inputs = tokenizer(prompt, return_tensors="pt")
39
+ outputs = model.generate(inputs["input_ids"])
40
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
41
+ ```
42
+
43
+ ## 📈 Evaluation and Benchmarks
44
+ PersianLLaMA demonstrates superior performance over existing models, with robust evaluation metrics that highlight its capabilities in natural language understanding and generation.
45
+
46
+
47
+ ## 📜 Citing PersianLLaMA
48
+ If you find PersianLLaMA useful in your research, please consider citing:
49
+
50
+ ```sql
51
+ @article{abbasi2023persianllama,
52
+ title={PersianLLaMA: Towards Building First Persian Large Language Model},
53
+ author={Abbasi, Mohammad Amin and others},
54
+ journal={https://arxiv.org/abs/2312.15713},
55
+ year={2023}
56
+ }
57
+ ```
58
+
59
+
60
+ ## 📄 License
61
+ PersianLLaMA is open-sourced under the MIT license.