satpalsr commited on
Commit
a12e336
1 Parent(s): e6d6663

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "satpalsr/gemma-7b",
3
  "architectures": [
4
  "GemmaForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "mhenrichsen/gemma-7b",
3
  "architectures": [
4
  "GemmaForCausalLM"
5
  ],
pytorch_model-00001-of-00004.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4222146a7346252cb2e3c995aabca0a9c501e3de3945bb354f66248fc8d0d7c2
3
  size 4995511092
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6de4a5bf09a21a45f2725c0c871bd6192bcf31e3b50b978c0d0b04ca32a50d97
3
  size 4995511092
pytorch_model-00002-of-00004.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4926a7494d24583fa12e178a7be89b91cb43bd7e05e98004c8684e014dc45e3e
3
  size 4982972660
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49a89e4442d013f0ac524d4f9fce5341b6fc97fad663157c7049b0227303f783
3
  size 4982972660
pytorch_model-00003-of-00004.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eec38a4b818124b520385f61df63c60ad49e951bb88f10ee36fd1383d8875742
3
  size 4982972724
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a11749b4efcb2d6d2d511e0e9114cfedb644ae6d385f1c9ea16b5eb54c87c7e
3
  size 4982972724
pytorch_model-00004-of-00004.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c625db115f7ec04dd0e39f25856a6fce5af084d712d1e518a567159959e8042f
3
  size 2113996412
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:589823a9c27096cb1fffe8cc6b766c9290c82cba930a769acd8d8867249cd182
3
  size 2113996412
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cc2b347eecc3da9f14157dfaf6775c6d325868e5afd16c1ee2153fc5c5a4c630
3
- size 17477917
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0d908b4f9326e0998815690e325b6abbd378978553e10627924dd825db7e243
3
+ size 17477553
tokenizer.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:61a7b147390c64585d6c3543dd6fc636906c9af3865a5548f27f31aee1d4c8e2
3
- size 4241003
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6969e64047744a44bb3abfb5c50f8de0f7ed8b571d5444426ef931f651d1a0ef
3
+ size 4241111
tokenizer_config.json CHANGED
@@ -33,26 +33,9 @@
33
  "rstrip": false,
34
  "single_word": false,
35
  "special": true
36
- },
37
- "106": {
38
- "content": "<|im_start|>",
39
- "lstrip": false,
40
- "normalized": false,
41
- "rstrip": false,
42
- "single_word": false,
43
- "special": true
44
- },
45
- "107": {
46
- "content": "<|im_end|>",
47
- "lstrip": false,
48
- "normalized": false,
49
- "rstrip": false,
50
- "single_word": false,
51
- "special": true
52
  }
53
  },
54
  "bos_token": "<bos>",
55
- "chat_template": "{% if messages[0]['role'] == 'user' or messages[0]['role'] == 'system' %}{{ bos_token }}{% endif %}{% for message in messages %}{{ '<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n' }}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% elif messages[-1]['role'] == 'assistant' %}{{ eos_token }}{% endif %}",
56
  "clean_up_tokenization_spaces": false,
57
  "eos_token": "<eos>",
58
  "legacy": null,
 
33
  "rstrip": false,
34
  "single_word": false,
35
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
36
  }
37
  },
38
  "bos_token": "<bos>",
 
39
  "clean_up_tokenization_spaces": false,
40
  "eos_token": "<eos>",
41
  "legacy": null,