bytetriper commited on
Commit
ea7bf03
1 Parent(s): ab68263

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -12
README.md CHANGED
@@ -1,11 +1,10 @@
1
- ---
2
- license: apache-2.0
3
- language:
4
- - en
5
- pipeline_tag: image-to-image
6
- datasets:
7
- - ILSVRC/imagenet-1k
8
- ---
9
  # Model Card for Model ID
10
  VIT-MAE-r is a fine-tuned version of MAE for image reconstuction. We release a version fine-tuned from [MAE-Large](https://huggingface.co/facebook/vit-mae-large)
11
 
@@ -17,8 +16,8 @@ VIT-MAE-r is already converted to hf format and should be able to be used direct
17
 
18
  <!-- Provide the basic links for the model. -->
19
 
20
- - **Repository:** [More Information Needed]
21
- - **Paper [optional]:** [LM4LV: A Frozen Large Language Model for Low-level Vision Tasks](https://arxiv.org/abs/2405.15734v1)
22
  - **source model**: [MAE-Large](https://huggingface.co/facebook/vit-mae-large)
23
 
24
  ## How to Get Started with the Model
@@ -35,7 +34,7 @@ model = AutoModelForPreTraining.from_pretrained("bytetriper/vit-mae-r")
35
 
36
  This model achieves a rFID on ImageNet val set of 1.24, evaluated using the standard tensorflow tool provided by [Guided-Diffusion](https://github.com/openai/guided-diffusion/tree/main/evaluations)
37
 
38
- ## Citation [optional]
39
 
40
  <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
41
 
@@ -50,7 +49,7 @@ This model achieves a rFID on ImageNet val set of 1.24, evaluated using the stan
50
 
51
 
52
 
53
- ## Model Card Authors [optional]
54
 
55
  Boyang Zheng
56
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ datasets:
6
+ - ILSVRC/imagenet-1k
7
+ ---
 
8
  # Model Card for Model ID
9
  VIT-MAE-r is a fine-tuned version of MAE for image reconstuction. We release a version fine-tuned from [MAE-Large](https://huggingface.co/facebook/vit-mae-large)
10
 
 
16
 
17
  <!-- Provide the basic links for the model. -->
18
 
19
+ - **Repository:** [LM4LV](https://github.com/bytetriper/LM4LV)
20
+ - **Paper:** [LM4LV: A Frozen Large Language Model for Low-level Vision Tasks](https://arxiv.org/abs/2405.15734v1)
21
  - **source model**: [MAE-Large](https://huggingface.co/facebook/vit-mae-large)
22
 
23
  ## How to Get Started with the Model
 
34
 
35
  This model achieves a rFID on ImageNet val set of 1.24, evaluated using the standard tensorflow tool provided by [Guided-Diffusion](https://github.com/openai/guided-diffusion/tree/main/evaluations)
36
 
37
+ ## Citation
38
 
39
  <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
40
 
 
49
 
50
 
51
 
52
+ ## Model Card Authors
53
 
54
  Boyang Zheng
55