DarshanDeshpande commited on
Commit
738115b
1 Parent(s): 5f274d4

gemma_2b_oasst1_reward_model

Browse files
Files changed (3) hide show
  1. README.md +5 -5
  2. adapter_config.json +3 -5
  3. adapter_model.safetensors +1 -1
README.md CHANGED
@@ -20,8 +20,8 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [google/gemma-2b](https://huggingface.co/google/gemma-2b) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 0.4250
24
- - Accuracy: 0.7881
25
 
26
  ## Model description
27
 
@@ -55,9 +55,9 @@ The following hyperparameters were used during training:
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
57
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
58
- | 0.443 | 1.0 | 100 | 0.5045 | 0.7458 |
59
- | 0.4098 | 2.0 | 200 | 0.4312 | 0.7938 |
60
- | 0.5036 | 2.99 | 300 | 0.4250 | 0.7881 |
61
 
62
 
63
  ### Framework versions
 
20
 
21
  This model is a fine-tuned version of [google/gemma-2b](https://huggingface.co/google/gemma-2b) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.4345
24
+ - Accuracy: 0.8051
25
 
26
  ## Model description
27
 
 
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
57
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
58
+ | 0.5106 | 1.0 | 100 | 0.5843 | 0.7203 |
59
+ | 0.4299 | 2.0 | 200 | 0.4418 | 0.7825 |
60
+ | 0.5035 | 2.99 | 300 | 0.4345 | 0.8051 |
61
 
62
 
63
  ### Framework versions
adapter_config.json CHANGED
@@ -13,16 +13,14 @@
13
  "lora_dropout": 0.1,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
- "modules_to_save": [
17
- "score"
18
- ],
19
  "peft_type": "LORA",
20
  "r": 16,
21
  "rank_pattern": {},
22
  "revision": null,
23
  "target_modules": [
24
- "v_proj",
25
- "q_proj"
26
  ],
27
  "task_type": "SEQ_CLS",
28
  "use_dora": false,
 
13
  "lora_dropout": 0.1,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
+ "modules_to_save": null,
 
 
17
  "peft_type": "LORA",
18
  "r": 16,
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "q_proj",
23
+ "v_proj"
24
  ],
25
  "task_type": "SEQ_CLS",
26
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6fbbe5c5a3cbdef631341ec6ca752168bcbfb155d7263b87eda0c2a0de80bc47
3
  size 7390624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1205a50454ca4ff96b48f390603c27297d03af0968627f322716ec074ecc9663
3
  size 7390624