mpalaval commited on
Commit
1dc99ab
1 Parent(s): 80ad88c

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,13 +1,17 @@
1
  {
2
- "_name_or_path": "distilbert-base-uncased",
3
  "activation": "gelu",
4
  "architectures": [
5
  "DistilBertForMultilabelSequenceClassification"
6
  ],
7
  "attention_dropout": 0.1,
 
8
  "dim": 768,
9
  "dropout": 0.1,
 
 
10
  "hidden_dim": 3072,
 
11
  "id2label": [
12
  {
13
  "0": "admiration",
@@ -41,6 +45,7 @@
41
  }
42
  ],
43
  "initializer_range": 0.02,
 
44
  "label2id": {
45
  "admiration": 0,
46
  "amusement": 1,
@@ -71,16 +76,19 @@
71
  "sadness": 25,
72
  "surprise": 26
73
  },
 
74
  "max_position_embeddings": 512,
75
  "model_type": "distilbert",
76
  "n_heads": 12,
77
- "n_layers": 6,
78
  "pad_token_id": 0,
 
79
  "qa_dropout": 0.1,
80
  "seq_classif_dropout": 0.2,
81
  "sinusoidal_pos_embds": false,
82
- "tie_weights_": true,
83
  "torch_dtype": "float32",
84
  "transformers_version": "4.35.1",
85
- "vocab_size": 30522
 
 
86
  }
 
1
  {
2
+ "_name_or_path": "bert-base-cased",
3
  "activation": "gelu",
4
  "architectures": [
5
  "DistilBertForMultilabelSequenceClassification"
6
  ],
7
  "attention_dropout": 0.1,
8
+ "attention_probs_dropout_prob": 0.1,
9
  "dim": 768,
10
  "dropout": 0.1,
11
+ "gradient_checkpointing": false,
12
+ "hidden_act": "gelu",
13
  "hidden_dim": 3072,
14
+ "hidden_dropout_prob": 0.1,
15
  "id2label": [
16
  {
17
  "0": "admiration",
 
45
  }
46
  ],
47
  "initializer_range": 0.02,
48
+ "intermediate_size": 3072,
49
  "label2id": {
50
  "admiration": 0,
51
  "amusement": 1,
 
76
  "sadness": 25,
77
  "surprise": 26
78
  },
79
+ "layer_norm_eps": 1e-12,
80
  "max_position_embeddings": 512,
81
  "model_type": "distilbert",
82
  "n_heads": 12,
83
+ "n_layers": 12,
84
  "pad_token_id": 0,
85
+ "position_embedding_type": "absolute",
86
  "qa_dropout": 0.1,
87
  "seq_classif_dropout": 0.2,
88
  "sinusoidal_pos_embds": false,
 
89
  "torch_dtype": "float32",
90
  "transformers_version": "4.35.1",
91
+ "type_vocab_size": 2,
92
+ "use_cache": true,
93
+ "vocab_size": 28996
94
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:11a9d489bdc72976ab51281fb7f571a4f12b02fe60ea1961ad6c6e2e16f7ce09
3
- size 267912544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4413b3e1e0140983503d6848ef6f8afd31b5597a28b3b0c7f83910eb5220ff19
3
+ size 433345360
runs/Nov15_02-58-57_8ef5647a71ed/events.out.tfevents.1700017176.8ef5647a71ed.1876.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f120204611f0717e0e06510e90521d8672bc165bdf35b77a873d6c491b6373e
3
+ size 6909
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -43,13 +43,13 @@
43
  },
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
- "do_lower_case": true,
47
  "mask_token": "[MASK]",
48
  "model_max_length": 512,
49
  "pad_token": "[PAD]",
50
  "sep_token": "[SEP]",
51
  "strip_accents": null,
52
  "tokenize_chinese_chars": true,
53
- "tokenizer_class": "DistilBertTokenizer",
54
  "unk_token": "[UNK]"
55
  }
 
43
  },
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
+ "do_lower_case": false,
47
  "mask_token": "[MASK]",
48
  "model_max_length": 512,
49
  "pad_token": "[PAD]",
50
  "sep_token": "[SEP]",
51
  "strip_accents": null,
52
  "tokenize_chinese_chars": true,
53
+ "tokenizer_class": "BertTokenizer",
54
  "unk_token": "[UNK]"
55
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e88732b27c565a9c2fb0bc09e57b159ef6bc0f0eff810fe65ad8661768bf6836
3
  size 4664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f942bfd1c3c866f40cfc4ca7b88dc43495f4475520ea98e222691ccf68e3d506
3
  size 4664
vocab.txt CHANGED
The diff for this file is too large to render. See raw diff