xshubhamx commited on
Commit
e94507a
1 Parent(s): b8d8c4b

Upload folder using huggingface_hub

Browse files
training_checkpoints/checkpoint-321/README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  library_name: peft
3
- base_model: facebook/bart-base
4
  ---
5
 
6
  # Model Card for Model ID
@@ -81,7 +81,7 @@ Use the code below to get started with the model.
81
 
82
  [More Information Needed]
83
 
84
- ### Training Procedure
85
 
86
  <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
 
@@ -197,8 +197,6 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
197
  ## Model Card Contact
198
 
199
  [More Information Needed]
200
-
201
-
202
  ### Framework versions
203
 
204
  - PEFT 0.10.0
 
1
  ---
2
  library_name: peft
3
+ base_model: facebook/bart-large
4
  ---
5
 
6
  # Model Card for Model ID
 
81
 
82
  [More Information Needed]
83
 
84
+ ### Training Procedure
85
 
86
  <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
 
 
197
  ## Model Card Contact
198
 
199
  [More Information Needed]
 
 
200
  ### Framework versions
201
 
202
  - PEFT 0.10.0
training_checkpoints/checkpoint-321/adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "facebook/bart-base",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "dense",
24
- "v_proj",
25
  "fc1",
26
- "fc2",
27
- "out_proj",
28
  "q_proj",
29
- "k_proj"
 
 
 
 
30
  ],
31
  "task_type": "SEQ_CLS",
32
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "facebook/bart-large",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
23
  "fc1",
 
 
24
  "q_proj",
25
+ "dense",
26
+ "k_proj",
27
+ "out_proj",
28
+ "fc2",
29
+ "v_proj"
30
  ],
31
  "task_type": "SEQ_CLS",
32
  "use_dora": false,
training_checkpoints/checkpoint-321/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cd0f7ad66abd890021dddea05703ef62191b8929563f51ea85805837fab2f99b
3
- size 13151448
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f090365d3da6b8284306a4ba3d76042c6b984b2b5ac27c43973b3cdfbb3754d8
3
+ size 34854640
training_checkpoints/checkpoint-321/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:816652182da3c17938db31179fbe9bb3b336ab7a2ab7b2fd8b6dfb8ae753243a
3
- size 26413626
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb91a5e912478bd42fd0d9555af08c8edd180d1251350c1e263d2238747c03ec
3
+ size 69927198
training_checkpoints/checkpoint-321/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aaf53fe1a8c71059a38da0c7e1555fa2e86ea713b625d24386a7fff396909278
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:089ca74c65c8493973c9ac84df53873d9b0b4b7233ef599ed53f08fdd44ff5ad
3
  size 14244
training_checkpoints/checkpoint-321/trainer_state.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "best_metric": 0.3947617401663212,
3
  "best_model_checkpoint": "bart-base-lora/checkpoint-321",
4
  "epoch": 1.9968895800933126,
5
  "eval_steps": 500,
@@ -10,46 +10,46 @@
10
  "log_history": [
11
  {
12
  "epoch": 1.0,
13
- "eval_accuracy": 0.611154144074361,
14
- "eval_f1_macro": 0.28705545756004525,
15
- "eval_f1_micro": 0.611154144074361,
16
- "eval_f1_weighted": 0.5574695078229231,
17
- "eval_loss": 1.3204560279846191,
18
- "eval_macro_fpr": 0.046390243375546196,
19
- "eval_macro_sensitivity": 0.302378554842108,
20
- "eval_macro_specificity": 0.9691815961629444,
21
- "eval_precision": 0.5321554731456403,
22
- "eval_precision_macro": 0.2887439395071009,
23
- "eval_recall": 0.611154144074361,
24
- "eval_recall_macro": 0.302378554842108,
25
- "eval_runtime": 45.1674,
26
- "eval_samples_per_second": 28.583,
27
- "eval_steps_per_second": 3.587,
28
- "eval_weighted_fpr": 0.04347073086248701,
29
- "eval_weighted_sensitivity": 0.611154144074361,
30
- "eval_weighted_specificity": 0.9265697983698051,
31
  "step": 160
32
  },
33
  {
34
  "epoch": 2.0,
35
- "eval_accuracy": 0.6994577846630519,
36
- "eval_f1_macro": 0.3947617401663212,
37
- "eval_f1_micro": 0.6994577846630519,
38
- "eval_f1_weighted": 0.6807748226338182,
39
- "eval_loss": 0.887535810470581,
40
- "eval_macro_fpr": 0.03062685085400977,
41
- "eval_macro_sensitivity": 0.4253981018601158,
42
- "eval_macro_specificity": 0.977359054380045,
43
- "eval_precision": 0.6728416067370143,
44
- "eval_precision_macro": 0.3822291983692299,
45
- "eval_recall": 0.6994577846630519,
46
- "eval_recall_macro": 0.4253981018601158,
47
- "eval_runtime": 44.0045,
48
- "eval_samples_per_second": 29.338,
49
- "eval_steps_per_second": 3.681,
50
- "eval_weighted_fpr": 0.02977743668457406,
51
- "eval_weighted_sensitivity": 0.6994577846630519,
52
- "eval_weighted_specificity": 0.9609280310376243,
53
  "step": 321
54
  }
55
  ],
@@ -57,7 +57,7 @@
57
  "max_steps": 2400,
58
  "num_train_epochs": 15,
59
  "save_steps": 500,
60
- "total_flos": 3255412271763456.0,
61
  "trial_name": null,
62
  "trial_params": null
63
  }
 
1
  {
2
+ "best_metric": 0.43370046411892155,
3
  "best_model_checkpoint": "bart-base-lora/checkpoint-321",
4
  "epoch": 1.9968895800933126,
5
  "eval_steps": 500,
 
10
  "log_history": [
11
  {
12
  "epoch": 1.0,
13
+ "eval_accuracy": 0.715724244771495,
14
+ "eval_f1_macro": 0.3957532686340496,
15
+ "eval_f1_micro": 0.715724244771495,
16
+ "eval_f1_weighted": 0.6834645556416842,
17
+ "eval_loss": 0.9525313973426819,
18
+ "eval_macro_fpr": 0.028501626692807605,
19
+ "eval_macro_sensitivity": 0.4415906678049911,
20
+ "eval_macro_specificity": 0.9786637894661117,
21
+ "eval_precision": 0.6787596592263883,
22
+ "eval_precision_macro": 0.38751339695744513,
23
+ "eval_recall": 0.715724244771495,
24
+ "eval_recall_macro": 0.4415906678049911,
25
+ "eval_runtime": 120.9594,
26
+ "eval_samples_per_second": 10.673,
27
+ "eval_steps_per_second": 1.339,
28
+ "eval_weighted_fpr": 0.027587762158911525,
29
+ "eval_weighted_sensitivity": 0.715724244771495,
30
+ "eval_weighted_specificity": 0.9642325972201805,
31
  "step": 160
32
  },
33
  {
34
  "epoch": 2.0,
35
+ "eval_accuracy": 0.7412858249419055,
36
+ "eval_f1_macro": 0.43370046411892155,
37
+ "eval_f1_micro": 0.7412858249419055,
38
+ "eval_f1_weighted": 0.7231329193382661,
39
+ "eval_loss": 0.7733433842658997,
40
+ "eval_macro_fpr": 0.025167370687133015,
41
+ "eval_macro_sensitivity": 0.4686708989603177,
42
+ "eval_macro_specificity": 0.9805374797463638,
43
+ "eval_precision": 0.729625861926521,
44
+ "eval_precision_macro": 0.4491323101141271,
45
+ "eval_recall": 0.7412858249419055,
46
+ "eval_recall_macro": 0.4686708989603177,
47
+ "eval_runtime": 119.8649,
48
+ "eval_samples_per_second": 10.77,
49
+ "eval_steps_per_second": 1.352,
50
+ "eval_weighted_fpr": 0.024322749781532187,
51
+ "eval_weighted_sensitivity": 0.7412858249419055,
52
+ "eval_weighted_specificity": 0.9667763712535515,
53
  "step": 321
54
  }
55
  ],
 
57
  "max_steps": 2400,
58
  "num_train_epochs": 15,
59
  "save_steps": 500,
60
+ "total_flos": 1.1442847472308224e+16,
61
  "trial_name": null,
62
  "trial_params": null
63
  }
training_checkpoints/checkpoint-321/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cfb230625284b71a8582d0495ae9a462b745f8cb976f9d7932c4f91357944609
3
  size 4600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cfeeb4079e0edbd60cad2df3b73a192a8e64f61f57cc52cfabb80fd9ff6105d
3
  size 4600