tanganke commited on
Commit
5221efd
1 Parent(s): 8fea6a6

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -5,20 +5,6 @@ library_name: peft
5
 
6
  # Model Card for clip-vit-base-patch16_svhn_l-lora-16
7
 
8
- ## Usage
9
-
10
- ```python
11
- from transformers import CLIPVisionModel
12
- from peft import PeftModel
13
- from fusion_bench.method.classification.clip_finetune import linearize_lora_model_
14
-
15
- base_model = CLIPVisionModel.from_pretrained("openai/clip-vit-base-patch16").vision_model
16
- peft_model_id = "openai/clip-vit-base-patch16_svhn_l-lora-16"
17
- model = PeftModel.from_pretrained(base_model, peft_model_id)
18
-
19
- model = linearize_lora_model_(model)
20
- ```
21
-
22
  ## Model Card Contact
23
 
24
  ### Framework versions
 
5
 
6
  # Model Card for clip-vit-base-patch16_svhn_l-lora-16
7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  ## Model Card Contact
9
 
10
  ### Framework versions
adapter_config.json CHANGED
@@ -1,13 +1,10 @@
1
  {
2
  "alpha_pattern": {},
3
- "auto_mapping": {
4
- "base_model_class": "CLIPVisionTransformer",
5
- "parent_library": "transformers.models.clip.modeling_clip"
6
- },
7
  "base_model_name_or_path": "openai/clip-vit-base-patch16",
8
  "bias": "none",
9
  "fan_in_fan_out": false,
10
- "inference_mode": true,
11
  "init_lora_weights": true,
12
  "layer_replication": null,
13
  "layers_pattern": null,
 
1
  {
2
  "alpha_pattern": {},
3
+ "auto_mapping": null,
 
 
 
4
  "base_model_name_or_path": "openai/clip-vit-base-patch16",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
+ "inference_mode": false,
8
  "init_lora_weights": true,
9
  "layer_replication": null,
10
  "layers_pattern": null,
adapter_model.safetensors → linearized_adapter_model.safetensors RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:23e85f3956f3eba6772ec4a00a98859a88d17324b1190a2ccf488a0a2893eee9
3
- size 2365680
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:117da1580cff4e5fa1f8d3fc905fda1c69433689d8543408402578dde22da000
3
+ size 4732152