harpomaxx commited on
Commit
7562039
1 Parent(s): 2920594

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -3
README.md CHANGED
@@ -11,16 +11,20 @@ library_name: adapter-transformers
11
 
12
  ## Model Description
13
 
14
- A simple opt350m model trained on the CodeAlpaca dataset using quantization and Progressive Embedding Fine-Tuning (PEFT). It's designed to understand and generate code-related responses based on the prompts provided.
 
 
 
15
 
16
  ### Model Architecture
17
 
18
  - **Base Model**: `facebook/opt-350m`
19
- - **Fine-tuning**: Progressive Embedding Fine-Tuning (PEFT)
20
 
21
  ## Training Data
22
 
23
  The model was trained on the `lucasmccabe-lmi/CodeAlpaca-20k` dataset. This dataset contains code-related prompts and their corresponding outputs.
 
24
 
25
  ## Training Procedure
26
 
@@ -61,7 +65,7 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
61
  tokenizer = AutoTokenizer.from_pretrained("facebook/opt350m")
62
  model = AutoModelForCausalLM.from_pretrained("harpomaxx/opt350m-codealpaca20k)
63
 
64
- prompt = "### Question: [Your code-related question here]"
65
  inputs = tokenizer.encode(prompt, return_tensors="pt")
66
  outputs = model.generate(inputs)
67
  decoded_output = tokenizer.decode(outputs[0], skip_special_tokens=True)
 
11
 
12
  ## Model Description
13
 
14
+ An opt-350m model trained on the CodeAlpaca 20k dataset using quantization and Progressive Embedding Fine-Tuning (PEFT).
15
+ The resulting model is designed to understand and generate code-related responses based on the prompts provided.
16
+
17
+ [original model car](https://huggingface.co/facebook/opt-350m)
18
 
19
  ### Model Architecture
20
 
21
  - **Base Model**: `facebook/opt-350m`
22
+ - **Fine-tuning**: Parameter-Efficient Fine-Tuning (PEFT)
23
 
24
  ## Training Data
25
 
26
  The model was trained on the `lucasmccabe-lmi/CodeAlpaca-20k` dataset. This dataset contains code-related prompts and their corresponding outputs.
27
+ Script used for training is avaiable [here](https://github.com/harpomaxx/llm-finetuning/blob/0954a7ca16bb25bdef6ee9dd1089867bd4d8e0a5/code/python/scripts/stf_train_opt350m.py)
28
 
29
  ## Training Procedure
30
 
 
65
  tokenizer = AutoTokenizer.from_pretrained("facebook/opt350m")
66
  model = AutoModelForCausalLM.from_pretrained("harpomaxx/opt350m-codealpaca20k)
67
 
68
+ prompt = "Question: [Your code-related question here] ### Answer: "
69
  inputs = tokenizer.encode(prompt, return_tensors="pt")
70
  outputs = model.generate(inputs)
71
  decoded_output = tokenizer.decode(outputs[0], skip_special_tokens=True)