divakaivan commited on
Commit
30490d0
1 Parent(s): bae2fee

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -2
README.md CHANGED
@@ -15,8 +15,13 @@ tags:
15
 
16
  ```python
17
  from unsloth import FastLanguageModel
 
 
 
 
 
18
  model, tokenizer = FastLanguageModel.from_pretrained(
19
- model_name = "divakaivan/llama3-finetuned-reviewer-kr", # YOUR MODEL YOU USED FOR TRAINING
20
  max_seq_length = max_seq_length,
21
  dtype = dtype,
22
  load_in_4bit = load_in_4bit,
@@ -26,7 +31,7 @@ FastLanguageModel.for_inference(model) # Enable native 2x faster inference
26
  inputs = tokenizer(
27
  [
28
  alpaca_prompt.format(
29
- "", # input
30
  "", # output - leave this blank for generation!
31
  )
32
  ], return_tensors = "pt").to("cuda")
 
15
 
16
  ```python
17
  from unsloth import FastLanguageModel
18
+
19
+ max_seq_length = 2048 # Choose any! We auto support RoPE Scaling internally!
20
+ dtype = None # None for auto detection. Float16 for Tesla T4, V100, Bfloat16 for Ampere+
21
+ load_in_4bit = True # Use 4bit quantization to reduce memory usage. Can be False.
22
+
23
  model, tokenizer = FastLanguageModel.from_pretrained(
24
+ model_name = "divakaivan/llama3-finetuned-reviewer-kr",
25
  max_seq_length = max_seq_length,
26
  dtype = dtype,
27
  load_in_4bit = load_in_4bit,
 
31
  inputs = tokenizer(
32
  [
33
  alpaca_prompt.format(
34
+ "", # input - place your input here
35
  "", # output - leave this blank for generation!
36
  )
37
  ], return_tensors = "pt").to("cuda")