--- license: openrail datasets: - lucasmccabe-lmi/CodeAlpaca-20k language: - en library_name: adapter-transformers --- # Model Card for `opt350m-codealpaca20k` ## Model Description A simple opt350m model trained on the CodeAlpaca dataset using quantization and Progressive Embedding Fine-Tuning (PEFT). It's designed to understand and generate code-related responses based on the prompts provided. ### Model Architecture - **Base Model**: `facebook/opt-350m` - **Fine-tuning**: Progressive Embedding Fine-Tuning (PEFT) ## Training Data The model was trained on the `lucasmccabe-lmi/CodeAlpaca-20k` dataset. This dataset contains code-related prompts and their corresponding outputs. ## Training Procedure ### Quantization Configuration: - **Quantization Type**: 4-bit - **Compute Dtype**: float16 - **Double Quant**: Enabled ### PEFT Configuration: - **Lora Alpha**: 16 - **Lora Dropout**: 0.5 - **Bias**: None - **Task Type**: CAUSAL_LM - **Target Modules**: q_proj, v_proj, k_proj ### Training Arguments: - **Output Directory**: `./results` - **Batch Size**: 4 (per device) - **Gradient Accumulation Steps**: 2 - **Number of Epochs**: 10 - **Optimizer**: `adamw_bnb_8bit` - **Learning Rate**: 2e-5 - **Max Gradient Norm**: 0.3 - **Warmup Ratio**: 0.03 - **Learning Rate Scheduler**: Cosine - **Logging Steps**: 10 - **Save Steps**: 250 - **FP16 Precision**: Enabled ## Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("facebook/opt350m") model = AutoModelForCausalLM.from_pretrained("harpomaxx/opt350m-codealpaca20k) prompt = "### Question: [Your code-related question here]" inputs = tokenizer.encode(prompt, return_tensors="pt") outputs = model.generate(inputs) decoded_output = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded_output) ``` ## License OpenRail ---