DarshanDeshpande commited on
Commit
342c126
1 Parent(s): b46a4cd

Push model using huggingface_hub.

Browse files
README.md CHANGED
@@ -25,7 +25,7 @@ You can then generate text as follows:
25
  ```python
26
  from transformers import pipeline
27
 
28
- generator = pipeline("text-generation", model="DarshanDeshpande//tmp/SLURM_20252742/tmp8za0h0al/DarshanDeshpande/gemma_2b_oasst1_ppo_model")
29
  outputs = generator("Hello, my llama is cute")
30
  ```
31
 
@@ -35,8 +35,8 @@ If you want to use the model for training or to obtain the outputs from the valu
35
  from transformers import AutoTokenizer
36
  from trl import AutoModelForCausalLMWithValueHead
37
 
38
- tokenizer = AutoTokenizer.from_pretrained("DarshanDeshpande//tmp/SLURM_20252742/tmp8za0h0al/DarshanDeshpande/gemma_2b_oasst1_ppo_model")
39
- model = AutoModelForCausalLMWithValueHead.from_pretrained("DarshanDeshpande//tmp/SLURM_20252742/tmp8za0h0al/DarshanDeshpande/gemma_2b_oasst1_ppo_model")
40
 
41
  inputs = tokenizer("Hello, my llama is cute", return_tensors="pt")
42
  outputs = model(**inputs, labels=inputs["input_ids"])
 
25
  ```python
26
  from transformers import pipeline
27
 
28
+ generator = pipeline("text-generation", model="DarshanDeshpande//tmp/SLURM_20253966/tmph2bl7syr/DarshanDeshpande/gemma_2b_oasst1_ppo_model")
29
  outputs = generator("Hello, my llama is cute")
30
  ```
31
 
 
35
  from transformers import AutoTokenizer
36
  from trl import AutoModelForCausalLMWithValueHead
37
 
38
+ tokenizer = AutoTokenizer.from_pretrained("DarshanDeshpande//tmp/SLURM_20253966/tmph2bl7syr/DarshanDeshpande/gemma_2b_oasst1_ppo_model")
39
+ model = AutoModelForCausalLMWithValueHead.from_pretrained("DarshanDeshpande//tmp/SLURM_20253966/tmph2bl7syr/DarshanDeshpande/gemma_2b_oasst1_ppo_model")
40
 
41
  inputs = tokenizer("Hello, my llama is cute", return_tensors="pt")
42
  outputs = model(**inputs, labels=inputs["input_ids"])
adapter_config.json CHANGED
@@ -19,8 +19,8 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "q_proj",
23
- "v_proj"
24
  ],
25
  "task_type": "CAUSAL_LM",
26
  "use_dora": false,
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "v_proj",
23
+ "q_proj"
24
  ],
25
  "task_type": "CAUSAL_LM",
26
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2690fbb25dd45c0050b9248acd6807fa880eabf4c656cf137058b1891a5e2f17
3
  size 3695848
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60130f2e757132e31170ed5b40cd440c9ecd5d2351a2accc897956f2951fb4f4
3
  size 3695848
config.json CHANGED
@@ -26,9 +26,9 @@
26
  "cliprange": 0.2,
27
  "cliprange_value": 0.2,
28
  "vf_coef": 0.1,
29
- "batch_size": 16,
30
  "forward_batch_size": null,
31
- "mini_batch_size": 16,
32
  "gradient_accumulation_steps": 1,
33
  "world_size": 1,
34
  "ppo_epochs": 1,
@@ -45,7 +45,7 @@
45
  "whiten_rewards": false,
46
  "is_encoder_decoder": false,
47
  "is_peft_model": true,
48
- "backward_batch_size": 16,
49
- "global_backward_batch_size": 16,
50
- "global_batch_size": 16
51
  }
 
26
  "cliprange": 0.2,
27
  "cliprange_value": 0.2,
28
  "vf_coef": 0.1,
29
+ "batch_size": 25,
30
  "forward_batch_size": null,
31
+ "mini_batch_size": 25,
32
  "gradient_accumulation_steps": 1,
33
  "world_size": 1,
34
  "ppo_epochs": 1,
 
45
  "whiten_rewards": false,
46
  "is_encoder_decoder": false,
47
  "is_peft_model": true,
48
+ "backward_batch_size": 25,
49
+ "global_backward_batch_size": 25,
50
+ "global_batch_size": 25
51
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dec772d4057432c43b26ec4d65232d7b68e31d3d3bc90c6d869ebda70fdf27f7
3
  size 9724
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4848be04eebc4fd14531016e726563a8291df25b605763e41bc037ec0b5e0ac9
3
  size 9724