cocoshe commited on
Commit
c307a35
1 Parent(s): 90c0a78
cogagent-chat/.lock/0e7d92e3cbc24276bfd4c65cd10882e9 ADDED
File without changes
cogagent-chat/.lock/1ef1bd908c870c9aea949ee362ebf9ea ADDED
File without changes
cogagent-chat/.lock/21fdaba1e39f7fd8a7b3d7bcb6610c72 ADDED
File without changes
cogagent-chat/.lock/3243ed50b31797d62080059ed931d69a ADDED
File without changes
cogagent-chat/.lock/3e9ffa822fdbeafe03adab5ecea51d73 ADDED
File without changes
cogagent-chat/.lock/aed22b8aafabc5def1f7026af03a4931 ADDED
File without changes
cogagent-chat/config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "cogagent",
3
+ "architectures": [
4
+ "CogModelForCausalLM"
5
+ ],
6
+ "auto_map": {
7
+ "AutoConfig": "configuration_cogagent.CogAgentConfig",
8
+ "AutoModelForCausalLM": "modeling_cogagent.CogAgentForCausalLM"
9
+ },
10
+ "model_type": "cogagent",
11
+ "bos_token_id": 1,
12
+ "cross_compute_hidden_size": 1024,
13
+ "cross_hidden_size": 1024,
14
+ "cross_image_size": 1120,
15
+ "eos_token_id": 2,
16
+ "hidden_act": "silu",
17
+ "hidden_size": 4096,
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 11008,
20
+ "max_position_embeddings": 2048,
21
+ "num_attention_heads": 32,
22
+ "num_hidden_layers": 32,
23
+ "pad_token_id": 0,
24
+ "rms_norm_eps": 1e-05,
25
+ "template_version": "chat",
26
+ "tie_word_embeddings": false,
27
+ "dtype": "float32",
28
+ "transformers_version": "4.36.0.dev0",
29
+ "use_cache": true,
30
+ "vision_config": {
31
+ "dropout_prob": 0.0,
32
+ "hidden_act": "gelu",
33
+ "hidden_size": 1792,
34
+ "image_size": 224,
35
+ "in_channels": 3,
36
+ "intermediate_size": 15360,
37
+ "layer_norm_eps": 1e-06,
38
+ "num_heads": 16,
39
+ "num_hidden_layers": 63,
40
+ "num_positions": 257,
41
+ "patch_size": 14
42
+ },
43
+ "vocab_size": 32000
44
+ }
cogagent-chat/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.36.0.dev0"
7
+ }
cogagent-chat.zip → cogagent-chat/model_state.pdparams RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b44d6adfa7df08ea08022aec2016f45a7d23966ce3a67d1fce33ab2bf4ba6a07
3
- size 28329570892
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c7c41e88a718df97dec7abf5bb061f0d664e7cb6cac7869f365204b88f55cb2
3
+ size 36623258300
cogvlm-chat.zip → cogagent-chat/sentencepiece.bpe.model RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0df25693093dfe773d3586ad214b2082c274af638da07f56b7434afc24cfe126
3
- size 27907838833
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
cogagent-chat/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<unk>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
cogagent-chat/tokenizer_config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "<s>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "clean_up_tokenization_spaces": false,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "</s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "legacy": false,
22
+ "model_max_length": 4096,
23
+ "pad_token": null,
24
+ "padding_side": "right",
25
+ "sp_model_kwargs": {},
26
+ "tokenizer_class": "LlamaTokenizer",
27
+ "unk_token": {
28
+ "__type": "AddedToken",
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false
34
+ }
35
+ }
cogvlm-chat/.lock/0fba76b42f139d51a80e71a522670d55 ADDED
File without changes
cogvlm-chat/.lock/74272f05e0a56aa83ef36b0f9286c9e0 ADDED
File without changes
cogvlm-chat/.lock/87ba7122dee19704acf274dbb0b953f2 ADDED
File without changes
cogvlm-chat/.lock/8c61b576551f1d8cfe84d69d67a23dca ADDED
File without changes
cogvlm-chat/.lock/e123ae8781e09975203d020be96c69c9 ADDED
File without changes
cogvlm-chat/.lock/e56525218b01907bfa2ffa855baa6682 ADDED
File without changes
cogvlm-chat/config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "cogvlm-chat-v1.1",
3
+ "architectures": [
4
+ "CogModelForCausalLM"
5
+ ],
6
+ "auto_map": {
7
+ "AutoConfig": "configuration_cogvlm.CogVLMConfig",
8
+ "AutoModelForCausalLM": "modeling_cogvlm.CogVLMForCausalLM"
9
+ },
10
+ "model_type": "cogvlm",
11
+ "bos_token_id": 1,
12
+ "eos_token_id": 2,
13
+ "hidden_act": "silu",
14
+ "hidden_size": 4096,
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 11008,
17
+ "max_position_embeddings": 2048,
18
+ "num_attention_heads": 32,
19
+ "num_hidden_layers": 32,
20
+ "pad_token_id": 0,
21
+ "rms_norm_eps": 1e-05,
22
+ "template_version": "chat",
23
+ "tie_word_embeddings": false,
24
+ "dtype": "float32",
25
+ "transformers_version": "4.35.0",
26
+ "use_cache": true,
27
+ "vision_config": {
28
+ "dropout_prob": 0.0,
29
+ "hidden_act": "gelu",
30
+ "hidden_size": 1792,
31
+ "image_size": 490,
32
+ "in_channels": 3,
33
+ "intermediate_size": 15360,
34
+ "layer_norm_eps": 1e-06,
35
+ "num_heads": 16,
36
+ "num_hidden_layers": 63,
37
+ "num_positions": 1226,
38
+ "patch_size": 14
39
+ },
40
+ "vocab_size": 32000
41
+ }
cogvlm-chat/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.35.0"
7
+ }
cogvlm-chat/model_state.pdparams ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c1911a20c9ff81604ce96dc7529d680672e7b2477f5330e76870127d61009d2
3
+ size 35954440478
cogvlm-chat/sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
cogvlm-chat/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<unk>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
cogvlm-chat/tokenizer_config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "<s>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "clean_up_tokenization_spaces": false,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "</s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "legacy": false,
22
+ "model_max_length": 4096,
23
+ "pad_token": null,
24
+ "padding_side": "right",
25
+ "sp_model_kwargs": {},
26
+ "tokenizer_class": "LlamaTokenizer",
27
+ "unk_token": {
28
+ "__type": "AddedToken",
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false
34
+ }
35
+ }