ani2857 commited on
Commit
c59e2f4
1 Parent(s): e67aa25

End of training

Browse files
Files changed (4) hide show
  1. README.md +1 -1
  2. pytorch_model.bin +2 -2
  3. tokenizer_config.json +0 -2
  4. training_args.bin +2 -2
README.md CHANGED
@@ -45,6 +45,6 @@ The following hyperparameters were used during training:
45
  ### Framework versions
46
 
47
  - Transformers 4.34.1
48
- - Pytorch 2.0.1+cu118
49
  - Datasets 2.14.5
50
  - Tokenizers 0.14.1
 
45
  ### Framework versions
46
 
47
  - Transformers 4.34.1
48
+ - Pytorch 2.1.0+cu118
49
  - Datasets 2.14.5
50
  - Tokenizers 0.14.1
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:32b1cb716e4e7c8234a22740772f8f4bae49dd8df85f13e3506be796fa04712e
3
- size 709122473
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc4dc53c1ebeafb5253a77ca4ed5fe6a22be92491d7d82552df1d19e2529fdd9
3
+ size 709125734
tokenizer_config.json CHANGED
@@ -43,11 +43,9 @@
43
  },
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
- "do_basic_tokenize": true,
47
  "do_lower_case": false,
48
  "mask_token": "[MASK]",
49
  "model_max_length": 512,
50
- "never_split": null,
51
  "pad_token": "[PAD]",
52
  "sep_token": "[SEP]",
53
  "strip_accents": null,
 
43
  },
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
 
46
  "do_lower_case": false,
47
  "mask_token": "[MASK]",
48
  "model_max_length": 512,
 
49
  "pad_token": "[PAD]",
50
  "sep_token": "[SEP]",
51
  "strip_accents": null,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3dca0878f817a23328f8e2dddcd2261006ced4dc5c2ccf8226e1cccbc99bb5d1
3
- size 4091
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d709f18384564fb00235ccde5f67b1ce783002a98d8d9684a5f4acae127287d
3
+ size 4536