ecker commited on
Commit
2d624f1
1 Parent(s): 67ce2fc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -1
README.md CHANGED
@@ -81,4 +81,23 @@ Some additional configurations have been explored with, but experiments have not
81
  Some current "achitectural features" are in-use, but their effects need to be experimented with further:
82
  * `split_classifier_heads` is still a mystery whether it's truly helpful or not (each RVQ level gets its own output head).
83
  * `audio_embeddings_sum` is also a mystery whether it matters if each later RVQ level should "see" the past levels through summing embeddings, or if not doing it is preferable.
84
- * Disabling `unified_position_ids` seems to help quality more often than not, but I'm still unsure if it's beneficial in practice.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
81
  Some current "achitectural features" are in-use, but their effects need to be experimented with further:
82
  * `split_classifier_heads` is still a mystery whether it's truly helpful or not (each RVQ level gets its own output head).
83
  * `audio_embeddings_sum` is also a mystery whether it matters if each later RVQ level should "see" the past levels through summing embeddings, or if not doing it is preferable.
84
+ * Disabling `unified_position_ids` seems to help quality more often than not, but I'm still unsure if it's beneficial in practice.
85
+
86
+ ## LoRAs
87
+
88
+ This repo also contains some LoRAs to serve as a reference under `./loras/`.
89
+
90
+ Using a LoRA is the same as a base model, except you're required to have the base model already (obviously). Just use the LoRA's config YAML to load from instead to use it.
91
+
92
+ The only caveat is that my original dataset *does* contain these samples already, but given the sheer size of it, they're probably underutilized.
93
+ * However, the base model already has *almost adequate* output from these speakers, but not enough to be satisfactory.
94
+
95
+ * `config.lora.glados.yaml` / `lora-glados-r128-a128`:
96
+ + A simple LoRA of GLaDOS from both Portal and Portal 2.
97
+ + Trained for 250 steps (48000 samples, 821 samples per epoch).
98
+ * `config.lora.sam.yaml` / `lora-sam-r128-a128`:
99
+ + A simple LoRA of Sam from the non-remaster Sam and Max Telltale games.
100
+ + Trained for 250 steps (48000 samples, 1555 samples per epoch).
101
+ * `config.lora.max.yaml` / `lora-max-r128-a128`:
102
+ + A simple LoRA of Max from the non-remaster Sam and Max Telltale games.
103
+ + Trained for 250 steps (48000 samples, 1292 samples per epoch).