srikanthsrnvs commited on
Commit
e679af4
1 Parent(s): b271591

Upload with huggingface_hub

Browse files
added_tokens.json CHANGED
@@ -1,3 +1,66 @@
1
  {
2
- "<pad>": 32000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  }
 
1
  {
2
+ "<pad>": 32000,
3
+ "<pad_0>": 32001,
4
+ "<pad_10>": 32011,
5
+ "<pad_11>": 32012,
6
+ "<pad_12>": 32013,
7
+ "<pad_13>": 32014,
8
+ "<pad_14>": 32015,
9
+ "<pad_15>": 32016,
10
+ "<pad_16>": 32017,
11
+ "<pad_17>": 32018,
12
+ "<pad_18>": 32019,
13
+ "<pad_19>": 32020,
14
+ "<pad_1>": 32002,
15
+ "<pad_20>": 32021,
16
+ "<pad_21>": 32022,
17
+ "<pad_22>": 32023,
18
+ "<pad_23>": 32024,
19
+ "<pad_24>": 32025,
20
+ "<pad_25>": 32026,
21
+ "<pad_26>": 32027,
22
+ "<pad_27>": 32028,
23
+ "<pad_28>": 32029,
24
+ "<pad_29>": 32030,
25
+ "<pad_2>": 32003,
26
+ "<pad_30>": 32031,
27
+ "<pad_31>": 32032,
28
+ "<pad_32>": 32033,
29
+ "<pad_33>": 32034,
30
+ "<pad_34>": 32035,
31
+ "<pad_35>": 32036,
32
+ "<pad_36>": 32037,
33
+ "<pad_37>": 32038,
34
+ "<pad_38>": 32039,
35
+ "<pad_39>": 32040,
36
+ "<pad_3>": 32004,
37
+ "<pad_40>": 32041,
38
+ "<pad_41>": 32042,
39
+ "<pad_42>": 32043,
40
+ "<pad_43>": 32044,
41
+ "<pad_44>": 32045,
42
+ "<pad_45>": 32046,
43
+ "<pad_46>": 32047,
44
+ "<pad_47>": 32048,
45
+ "<pad_48>": 32049,
46
+ "<pad_49>": 32050,
47
+ "<pad_4>": 32005,
48
+ "<pad_50>": 32051,
49
+ "<pad_51>": 32052,
50
+ "<pad_52>": 32053,
51
+ "<pad_53>": 32054,
52
+ "<pad_54>": 32055,
53
+ "<pad_55>": 32056,
54
+ "<pad_56>": 32057,
55
+ "<pad_57>": 32058,
56
+ "<pad_58>": 32059,
57
+ "<pad_59>": 32060,
58
+ "<pad_5>": 32006,
59
+ "<pad_60>": 32061,
60
+ "<pad_61>": 32062,
61
+ "<pad_62>": 32063,
62
+ "<pad_6>": 32007,
63
+ "<pad_7>": 32008,
64
+ "<pad_8>": 32009,
65
+ "<pad_9>": 32010
66
  }
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "swype/deepshard-13B-ft",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -17,7 +17,7 @@
17
  "rms_norm_eps": 1e-06,
18
  "tie_word_embeddings": false,
19
  "torch_dtype": "float32",
20
- "transformers_version": "4.27.0.dev0",
21
  "use_cache": true,
22
- "vocab_size": 32001
23
  }
 
1
  {
2
+ "_name_or_path": "swype/deepshard-13B-raw",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
17
  "rms_norm_eps": 1e-06,
18
  "tie_word_embeddings": false,
19
  "torch_dtype": "float32",
20
+ "transformers_version": "4.28.0.dev0",
21
  "use_cache": true,
22
+ "vocab_size": 32064
23
  }
generation_config.json CHANGED
@@ -3,5 +3,5 @@
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
  "pad_token_id": 0,
6
- "transformers_version": "4.27.0.dev0"
7
  }
 
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
  "pad_token_id": 0,
6
+ "transformers_version": "4.28.0.dev0"
7
  }
pytorch_model-00001-of-00006.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a149a0101fc8703e3dc6bbb53cab5cf4f1172e4947c7708af3190e2c750347e
3
- size 9956564363
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80f28164123d4cba241f975d8e83da164a8a876f201b2b5e5f590e3dd96c5d67
3
+ size 9957854603
pytorch_model-00002-of-00006.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b634d21c539b2e50ccec576a220135d230399ca931664788914f18137e9ae487
3
  size 9940856385
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:981a515a78f88ff17802cc8d3651176ec319e30d04723e64661621f85dc63519
3
  size 9940856385
pytorch_model-00003-of-00006.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5fbc9d6d5e12789561afaf7477e68a5910e9c4fe046c82be94a302f6866a2157
3
  size 9940856943
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:209f283e1486cf25fe3c7538640b9ce2664bc4c98d46ab2aeb463ba38bab862a
3
  size 9940856943
pytorch_model-00004-of-00006.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:69d3904a3c093a29d4f7daed9360ef259bc482578fe245aa535d8c7609e45885
3
  size 9867415289
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71c7556f26d95e7ecc1921f65f0d3103b276a9e386a88c3f473543b7b74b98aa
3
  size 9867415289
pytorch_model-00005-of-00006.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b194590a69eff170023132e3de4cdca5701a49d944ecb280a3c60fdfee4edd51
3
  size 9867456961
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fda681591cbec29c9ec2535ea29598f7f73e29824109e4540cf727fb54fcda02
3
  size 9867456961
pytorch_model-00006-of-00006.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:de508c13017859e80daefa6eb178ac0eca9a59145a794cb5b89ac8f9153304f7
3
- size 2490496687
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4192ca8d481e96c04c874a4066568ccc8301dff9cb533db41faf9e9df0cdb381
3
+ size 2491786927
pytorch_model.bin.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 52063508480
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "pytorch_model-00006-of-00006.bin",
 
1
  {
2
  "metadata": {
3
+ "total_size": 52066088960
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "pytorch_model-00006-of-00006.bin",
special_tokens_map.json CHANGED
@@ -1,3 +1,68 @@
1
  {
2
- "pad_token": "<pad>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  }
 
1
  {
2
+ "additional_special_tokens": [
3
+ "<pad_0>",
4
+ "<pad_1>",
5
+ "<pad_2>",
6
+ "<pad_3>",
7
+ "<pad_4>",
8
+ "<pad_5>",
9
+ "<pad_6>",
10
+ "<pad_7>",
11
+ "<pad_8>",
12
+ "<pad_9>",
13
+ "<pad_10>",
14
+ "<pad_11>",
15
+ "<pad_12>",
16
+ "<pad_13>",
17
+ "<pad_14>",
18
+ "<pad_15>",
19
+ "<pad_16>",
20
+ "<pad_17>",
21
+ "<pad_18>",
22
+ "<pad_19>",
23
+ "<pad_20>",
24
+ "<pad_21>",
25
+ "<pad_22>",
26
+ "<pad_23>",
27
+ "<pad_24>",
28
+ "<pad_25>",
29
+ "<pad_26>",
30
+ "<pad_27>",
31
+ "<pad_28>",
32
+ "<pad_29>",
33
+ "<pad_30>",
34
+ "<pad_31>",
35
+ "<pad_32>",
36
+ "<pad_33>",
37
+ "<pad_34>",
38
+ "<pad_35>",
39
+ "<pad_36>",
40
+ "<pad_37>",
41
+ "<pad_38>",
42
+ "<pad_39>",
43
+ "<pad_40>",
44
+ "<pad_41>",
45
+ "<pad_42>",
46
+ "<pad_43>",
47
+ "<pad_44>",
48
+ "<pad_45>",
49
+ "<pad_46>",
50
+ "<pad_47>",
51
+ "<pad_48>",
52
+ "<pad_49>",
53
+ "<pad_50>",
54
+ "<pad_51>",
55
+ "<pad_52>",
56
+ "<pad_53>",
57
+ "<pad_54>",
58
+ "<pad_55>",
59
+ "<pad_56>",
60
+ "<pad_57>",
61
+ "<pad_58>",
62
+ "<pad_59>",
63
+ "<pad_60>",
64
+ "<pad_61>",
65
+ "<pad_62>"
66
+ ],
67
+ "pad_token": "<pad_0>"
68
  }
tokenizer_config.json CHANGED
@@ -3,7 +3,7 @@
3
  "eos_token": "",
4
  "model_max_length": 512,
5
  "padding_side": "right",
6
- "special_tokens_map_file": "/home/paperspace/.cache/huggingface/hub/models--swype--deepshard-13B-raw/snapshots/3d3a439e49a5a7d3fb1a5177cecba2c1d785a7f4/special_tokens_map.json",
7
  "tokenizer_class": "LlamaTokenizer",
8
  "unk_token": ""
9
  }
 
3
  "eos_token": "",
4
  "model_max_length": 512,
5
  "padding_side": "right",
6
+ "special_tokens_map_file": "/home/ubuntu/.cache/huggingface/hub/models--swype--deepshard-13B-raw/snapshots/225c31ea4b7db4ec9277178b37bb8fbea50aa258/special_tokens_map.json",
7
  "tokenizer_class": "LlamaTokenizer",
8
  "unk_token": ""
9
  }
trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4d4da8ae412b13139c6c035d9a6eb6fd3381df3a217d9ca2b2e025b1d3ff0309
3
  size 3771
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08e19f9edc1c965b0c1203d81fb8b6a892916ee5857078395b7eeac813703eba
3
  size 3771