fragata commited on
Commit
f3bcbda
1 Parent(s): b286193

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +60 -0
README.md CHANGED
@@ -1,3 +1,63 @@
1
  ---
 
 
 
 
 
2
  license: cc-by-nc-4.0
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+
3
+ language:
4
+ - hu
5
+ tags:
6
+ - text-generation
7
  license: cc-by-nc-4.0
8
+ widget:
9
+ - text: "Elmesélek egy történetet a nyelvtechnológiáról."
10
  ---
11
+
12
+ # PULI GPT-3SX
13
+
14
+ For further details, see [our demo site](https://juniper.nytud.hu/demo/gpt2).
15
+
16
+ - Hungarian GPT-2 model
17
+ - Trained with Megatron-DeepSpeed [github](https://github.com/microsoft/Megatron-DeepSpeed)
18
+ - Dataset: 36.3 billion words
19
+ - Checkpoint: 500 000 steps
20
+
21
+ ## Limitations
22
+
23
+ - max_seq_length = 1024
24
+
25
+
26
+ ## Citation
27
+ If you use this model, please cite the following paper:
28
+
29
+ ```
30
+ @inproceedings {yang-gpt3,
31
+ title = {Jönnek a nagyok! GPT-3, GPT-2 és BERT large nyelvmodellek magyar nyelvre},
32
+ booktitle = {XIX. Magyar Számítógépes Nyelvészeti Konferencia (MSZNY 2023)},
33
+ year = {2023},
34
+ publisher = {Szegedi Tudományegyetem},
35
+ address = {Szeged, Hungary},
36
+ author = {Yang, Zijian Győző},
37
+ pages = {0}
38
+ }
39
+
40
+ ```
41
+
42
+ ## Usage
43
+
44
+ ```python
45
+ from transformers import GPT2Tokenizer, GPT2Model
46
+
47
+ tokenizer = GPT2Tokenizer.from_pretrained('NYTK/PULI-GPT-2')
48
+ model = GPT2Model.from_pretrained('NYTK/PULI-GPT-2')
49
+ text = "Replace me by any text you'd like."
50
+ encoded_input = tokenizer(text, return_tensors='pt')
51
+ output = model(**encoded_input)
52
+
53
+ ```
54
+ ## Usage with pipeline
55
+
56
+ ```python
57
+ from transformers import pipeline
58
+
59
+ prompt = "Elmesélek egy történetet a nyelvtechnológiáról."
60
+ generator = pipeline(task="text-generation", model="NYTK/PULI-GPT-3SX")
61
+
62
+ print(generator(prompt)[0]["generated_text"])
63
+ ```