jxie commited on
Commit
e44c65c
1 Parent(s): e59fd51

Upload SMAForSSL

Browse files
Files changed (1) hide show
  1. config.json +3 -13
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/iris/u/jwxie/workspace/domain-agnostic-pretraining/examples/research_projects/domain-agnostic-pretraining/saved_models/language_pretrained/wikibooks_guided_self_random_select_masking_recon_small-adamw_torch-lr1e-4-wd0.01-mr0.15/checkpoint-1000000",
3
  "architectures": [
4
  "SMAForSSL"
5
  ],
@@ -37,7 +37,7 @@
37
  "layernorm_eps": 1e-12,
38
  "loss_fn": "mse",
39
  "max_position_embeddings": 1024,
40
- "model_type": "perceiver_sma",
41
  "num_blocks": 1,
42
  "num_cross_attention_heads": 8,
43
  "num_discrete_tokens": 262,
@@ -60,17 +60,7 @@
60
  "ema_decay_end": 0.0,
61
  "ema_decay_start": 0.0
62
  },
63
- "eval_transform_args": [
64
- [
65
- "RandomlySelectedCrossAttentionMasking",
66
- {
67
- "exclude_seen_reconstruction": true,
68
- "masking_ratio": 0.15,
69
- "num_per_query": 4,
70
- "varying_length": true
71
- }
72
- ]
73
- ],
74
  "mask_replace": 3,
75
  "num_layer_target_avg": null,
76
  "reconstruction_decoder_args": {
 
1
  {
2
+ "_name_or_path": null,
3
  "architectures": [
4
  "SMAForSSL"
5
  ],
 
37
  "layernorm_eps": 1e-12,
38
  "loss_fn": "mse",
39
  "max_position_embeddings": 1024,
40
+ "model_type": "sma",
41
  "num_blocks": 1,
42
  "num_cross_attention_heads": 8,
43
  "num_discrete_tokens": 262,
 
60
  "ema_decay_end": 0.0,
61
  "ema_decay_start": 0.0
62
  },
63
+ "eval_transform_args": null,
 
 
 
 
 
 
 
 
 
 
64
  "mask_replace": 3,
65
  "num_layer_target_avg": null,
66
  "reconstruction_decoder_args": {