Edit model card

Model Card

We provide a low-rank adapter for an instruction-tuned 12B-parameter GPT3-style language model.

Prompting

For instructions that do not require extraneous inputs, the recommended prompt is:

Below is an instruction that describes a task. Write a response that appropriately completes the request.


### Instruction:

< -- instruction goes here --- >

### Response:

For instructions that do require extraneous inputs, the recommended prompt is:

Below is an instruction that describes a task. Write a response that appropriately completes the request.


### Instruction:

< -- instruction goes here -- >

### Input:

< -- extraneous input goes here -- >

### Response:

Since the model performs causal language modeling, the model's response to the prompt is the text completing the sequence beginning with the prompt.

Instruction-Tuning

This model was instruction-tuned on an internally-curated Alpaca-style dataset.

  • Epochs: 3
  • Batch size: 128
  • Cutoff length: 2048
  • Learning rate: 3.2e-5, cosine decay
  • LoRA r: 8
  • LoRA alpha: 16
  • LoRA dropout: 0.0
  • LoRA target modules: 'query_key_value', 'dense_h_to_4h', 'dense_4h_to_h', 'dense'
  • License: The instruction-tuning data is subject to the Creative Commons 4.0 license.

Base Model

This model was instruction-tuned from a 12B variant from the Pythia family.

Licensing Information

We release this adapter under the Creative Commons NonCommercial (CC BY-NC 4.0) license.

Downloads last month
8
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.