ecker commited on
Commit
67ce2fc
1 Parent(s): 5e1ff4a

Upload 8 files

Browse files
.gitattributes CHANGED
@@ -7,3 +7,6 @@ models/experiments/ckpt/nar-len-llama-9/fp32.pth filter=lfs diff=lfs merge=lfs -
7
  models/ckpt/ar+nar-llama-8/fp32.sft filter=lfs diff=lfs merge=lfs -text
8
  models/ckpt/ar+nar-retnet-8/fp32.sft filter=lfs diff=lfs merge=lfs -text
9
  models/ckpt/ar+nar-tts+stt-llama-8/fp32.sft filter=lfs diff=lfs merge=lfs -text
 
 
 
 
7
  models/ckpt/ar+nar-llama-8/fp32.sft filter=lfs diff=lfs merge=lfs -text
8
  models/ckpt/ar+nar-retnet-8/fp32.sft filter=lfs diff=lfs merge=lfs -text
9
  models/ckpt/ar+nar-tts+stt-llama-8/fp32.sft filter=lfs diff=lfs merge=lfs -text
10
+ loras/ckpt/lora-glados-r128-a128/lora.sft filter=lfs diff=lfs merge=lfs -text
11
+ loras/ckpt/lora-max-r128-a128/lora.sft filter=lfs diff=lfs merge=lfs -text
12
+ loras/ckpt/lora-sam-r128-a128/lora.sft filter=lfs diff=lfs merge=lfs -text
loras/ckpt/lora-glados-r128-a128/lora.sft ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e9bec9574b8a074c9a87f8997b27706f4f3611efa12d26dc9763cd2bca9eb9e
3
+ size 66076312
loras/ckpt/lora-max-r128-a128/lora.sft ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e8065b17f981dba303af2e5c92275902a45e8e38e9eea578f0a9e4b50690f1e
3
+ size 66076312
loras/ckpt/lora-sam-r128-a128/lora.sft ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:433ada1d6f9cba5ebce7917bb51df09f050f4d221afdc4cdad72964b6624f9fc
3
+ size 66076312
loras/config.lora.glados.yaml ADDED
@@ -0,0 +1,130 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ sample_rate: 24_000
2
+ audio_backend: "vocos"
3
+
4
+ models:
5
+ - name: "ar+nar-tts+stt"
6
+ size: "full"
7
+ resp_levels: 8
8
+ prom_levels: 8
9
+ tasks: 9
10
+ langs: 2
11
+ tones: 1
12
+ arch_type: llama
13
+ training: False
14
+ version: 5
15
+ attention: auto
16
+ dropout: 0.1
17
+ #loss_factors:
18
+ # text: 0.01
19
+ # prom: 0.5
20
+ # resp: 1.0
21
+ capabilities: ["ar", "nar"]
22
+ experimental:
23
+ p_rvq_levels: "auto"
24
+ audio_embedding_sums: True
25
+ unified_position_ids: False
26
+ split_classifiers: True
27
+ #
28
+ causal_size: 1
29
+ interleave: False
30
+ rvq_level_range: []
31
+ tie_classifier_to_embedding: False
32
+
33
+ loras:
34
+ - name : "lora-glados"
35
+ rank: 128
36
+ alpha: 128
37
+ training: True
38
+ rvq_levels: []
39
+
40
+ hyperparameters:
41
+ batch_size: 32
42
+ gradient_accumulation_steps: 8
43
+ gradient_clipping: 1.0
44
+ warmup_steps: 10
45
+
46
+ optimizer: Prodigy
47
+ learning_rate: 1.0
48
+ torch_optimizer: True
49
+
50
+ scheduler: "" # ScheduleFree
51
+ torch_scheduler: True
52
+
53
+ evaluation:
54
+ batch_size: 4
55
+ frequency: 250
56
+ size: 4
57
+
58
+ steps: 500
59
+ ar_temperature: 1.0
60
+ nar_temperature: 0.0
61
+
62
+ trainer:
63
+ iterations: 1_000_000
64
+ save_frequency: 250
65
+ keep_last_checkpoints: 4
66
+
67
+ resize_modules: True
68
+
69
+ check_for_oom: False
70
+ gradient_checkpointing: True
71
+
72
+ weight_dtype: bfloat16
73
+ amp: True
74
+
75
+ backend: deepspeed
76
+ deepspeed:
77
+ inferencing: False
78
+ amp: False
79
+
80
+ load_webui: False
81
+
82
+ inference:
83
+ backend: local
84
+ normalize: False
85
+
86
+ weight_dtype: bfloat16
87
+ amp: True
88
+
89
+ optimizations:
90
+ injects: False
91
+ replace: True
92
+
93
+ linear: False
94
+ embedding: False
95
+ optimizers: True
96
+
97
+ bitsandbytes: False
98
+ dadaptation: False
99
+ bitnet: False
100
+ fp8: False
101
+
102
+ dataset:
103
+ use_hdf5: True
104
+ hdf5_flag: r
105
+
106
+ use_metadata: True
107
+ validate: True
108
+
109
+ workers: 1
110
+ cache: True
111
+
112
+ duration_range: [3.0, 12.0]
113
+
114
+ random_utterance: 1.0
115
+ max_prompts: 1
116
+ prompt_duration_range: [3.0, 3.0]
117
+
118
+ max_resps: 1
119
+ p_resp_append: 0.25
120
+
121
+ sample_type: path # path # speaker
122
+ sample_order: duration
123
+ sample_max_duration_batch: 300
124
+ sample_shuffle: False
125
+
126
+ tasks_list: [ "tts", "stt" ]
127
+
128
+ training: []
129
+ validation: []
130
+ noise: []
loras/config.lora.max.yaml ADDED
@@ -0,0 +1,130 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ sample_rate: 24_000
2
+ audio_backend: "vocos"
3
+
4
+ models:
5
+ - name: "ar+nar-tts+stt"
6
+ size: "full"
7
+ resp_levels: 8
8
+ prom_levels: 8
9
+ tasks: 9
10
+ langs: 2
11
+ tones: 1
12
+ arch_type: llama
13
+ training: False
14
+ version: 5
15
+ attention: auto
16
+ dropout: 0.1
17
+ #loss_factors:
18
+ # text: 0.01
19
+ # prom: 0.5
20
+ # resp: 1.0
21
+ capabilities: ["ar", "nar"]
22
+ experimental:
23
+ p_rvq_levels: "auto"
24
+ audio_embedding_sums: True
25
+ unified_position_ids: False
26
+ split_classifiers: True
27
+ #
28
+ causal_size: 1
29
+ interleave: False
30
+ rvq_level_range: []
31
+ tie_classifier_to_embedding: False
32
+
33
+ loras:
34
+ - name : "lora-max"
35
+ rank: 128
36
+ alpha: 128
37
+ training: True
38
+ rvq_levels: []
39
+
40
+ hyperparameters:
41
+ batch_size: 32
42
+ gradient_accumulation_steps: 8
43
+ gradient_clipping: 1.0
44
+ warmup_steps: 10
45
+
46
+ optimizer: Prodigy
47
+ learning_rate: 1.0
48
+ torch_optimizer: True
49
+
50
+ scheduler: "" # ScheduleFree
51
+ torch_scheduler: True
52
+
53
+ evaluation:
54
+ batch_size: 4
55
+ frequency: 250
56
+ size: 4
57
+
58
+ steps: 500
59
+ ar_temperature: 1.0
60
+ nar_temperature: 0.0
61
+
62
+ trainer:
63
+ iterations: 1_000_000
64
+ save_frequency: 250
65
+ keep_last_checkpoints: 4
66
+
67
+ resize_modules: True
68
+
69
+ check_for_oom: False
70
+ gradient_checkpointing: True
71
+
72
+ weight_dtype: bfloat16
73
+ amp: True
74
+
75
+ backend: deepspeed
76
+ deepspeed:
77
+ inferencing: False
78
+ amp: False
79
+
80
+ load_webui: False
81
+
82
+ inference:
83
+ backend: local
84
+ normalize: False
85
+
86
+ weight_dtype: bfloat16
87
+ amp: True
88
+
89
+ optimizations:
90
+ injects: False
91
+ replace: True
92
+
93
+ linear: False
94
+ embedding: False
95
+ optimizers: True
96
+
97
+ bitsandbytes: False
98
+ dadaptation: False
99
+ bitnet: False
100
+ fp8: False
101
+
102
+ dataset:
103
+ use_hdf5: True
104
+ hdf5_flag: r
105
+
106
+ use_metadata: True
107
+ validate: True
108
+
109
+ workers: 1
110
+ cache: True
111
+
112
+ duration_range: [3.0, 12.0]
113
+
114
+ random_utterance: 1.0
115
+ max_prompts: 1
116
+ prompt_duration_range: [3.0, 3.0]
117
+
118
+ max_resps: 1
119
+ p_resp_append: 0.25
120
+
121
+ sample_type: path # path # speaker
122
+ sample_order: duration
123
+ sample_max_duration_batch: 300
124
+ sample_shuffle: False
125
+
126
+ tasks_list: [ "tts", "stt" ]
127
+
128
+ training: []
129
+ validation: []
130
+ noise: []
loras/config.lora.sam.yaml ADDED
@@ -0,0 +1,130 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ sample_rate: 24_000
2
+ audio_backend: "vocos"
3
+
4
+ models:
5
+ - name: "ar+nar-tts+stt"
6
+ size: "full"
7
+ resp_levels: 8
8
+ prom_levels: 8
9
+ tasks: 9
10
+ langs: 2
11
+ tones: 1
12
+ arch_type: llama
13
+ training: False
14
+ version: 5
15
+ attention: auto
16
+ dropout: 0.1
17
+ #loss_factors:
18
+ # text: 0.01
19
+ # prom: 0.5
20
+ # resp: 1.0
21
+ capabilities: ["ar", "nar"]
22
+ experimental:
23
+ p_rvq_levels: "auto"
24
+ audio_embedding_sums: True
25
+ unified_position_ids: False
26
+ split_classifiers: True
27
+ #
28
+ causal_size: 1
29
+ interleave: False
30
+ rvq_level_range: []
31
+ tie_classifier_to_embedding: False
32
+
33
+ loras:
34
+ - name : "lora-sam"
35
+ rank: 128
36
+ alpha: 128
37
+ training: True
38
+ rvq_levels: []
39
+
40
+ hyperparameters:
41
+ batch_size: 32
42
+ gradient_accumulation_steps: 8
43
+ gradient_clipping: 1.0
44
+ warmup_steps: 10
45
+
46
+ optimizer: Prodigy
47
+ learning_rate: 1.0
48
+ torch_optimizer: True
49
+
50
+ scheduler: "" # ScheduleFree
51
+ torch_scheduler: True
52
+
53
+ evaluation:
54
+ batch_size: 4
55
+ frequency: 250
56
+ size: 4
57
+
58
+ steps: 500
59
+ ar_temperature: 1.0
60
+ nar_temperature: 0.0
61
+
62
+ trainer:
63
+ iterations: 1_000_000
64
+ save_frequency: 250
65
+ keep_last_checkpoints: 4
66
+
67
+ resize_modules: True
68
+
69
+ check_for_oom: False
70
+ gradient_checkpointing: True
71
+
72
+ weight_dtype: bfloat16
73
+ amp: True
74
+
75
+ backend: deepspeed
76
+ deepspeed:
77
+ inferencing: False
78
+ amp: False
79
+
80
+ load_webui: False
81
+
82
+ inference:
83
+ backend: local
84
+ normalize: False
85
+
86
+ weight_dtype: bfloat16
87
+ amp: True
88
+
89
+ optimizations:
90
+ injects: False
91
+ replace: True
92
+
93
+ linear: False
94
+ embedding: False
95
+ optimizers: True
96
+
97
+ bitsandbytes: False
98
+ dadaptation: False
99
+ bitnet: False
100
+ fp8: False
101
+
102
+ dataset:
103
+ use_hdf5: True
104
+ hdf5_flag: r
105
+
106
+ use_metadata: True
107
+ validate: True
108
+
109
+ workers: 1
110
+ cache: True
111
+
112
+ duration_range: [3.0, 12.0]
113
+
114
+ random_utterance: 1.0
115
+ max_prompts: 1
116
+ prompt_duration_range: [3.0, 3.0]
117
+
118
+ max_resps: 1
119
+ p_resp_append: 0.25
120
+
121
+ sample_type: path # path # speaker
122
+ sample_order: duration
123
+ sample_max_duration_batch: 300
124
+ sample_shuffle: False
125
+
126
+ tasks_list: [ "tts", "stt" ]
127
+
128
+ training: []
129
+ validation: []
130
+ noise: []
models/ckpt/ar+nar-tts+stt-llama-8/fp32.sft CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cbf25cb391bbbb79ee05b5ecb9b219dce8785d9561473001e0e041a70d0e634b
3
  size 456272602
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc995fd0c81f7540518cd41a2c0482eb81f03a86c5d8f58dfd4aca4ec5fead3d
3
  size 456272602
models/config.llama.tts+stt.yaml CHANGED
@@ -20,14 +20,15 @@ models:
20
  # resp: 1.0
21
  capabilities: ["ar", "nar"]
22
  experimental:
23
- audio_embedding_sums: False
24
- interleave: False
25
  unified_position_ids: False
26
- rvq_level_range: []
27
  split_classifiers: True
28
- tie_classifier_to_embedding: False
29
  causal_size: 1
30
- p_rvq_levels: "auto"
 
 
31
 
32
  #loras:
33
  #- name : "lora"
@@ -99,13 +100,6 @@ optimizations:
99
  fp8: False
100
 
101
  dataset:
102
- speaker_name_getter: "lambda p: f'{p.parts[-3]}_{p.parts[-2]}'"
103
- speaker_group_getter: "lambda p: f'{p.parts[-3]}'"
104
- speaker_languages:
105
- ja: [
106
- "housamo"
107
- ]
108
-
109
  use_hdf5: True
110
  hdf5_flag: r
111
 
@@ -129,6 +123,8 @@ dataset:
129
  sample_max_duration_batch: 300
130
  sample_shuffle: False
131
 
 
 
132
  training: []
133
  validation: []
134
  noise: []
 
20
  # resp: 1.0
21
  capabilities: ["ar", "nar"]
22
  experimental:
23
+ p_rvq_levels: "auto"
24
+ audio_embedding_sums: True
25
  unified_position_ids: False
 
26
  split_classifiers: True
27
+ #
28
  causal_size: 1
29
+ interleave: False
30
+ rvq_level_range: []
31
+ tie_classifier_to_embedding: False
32
 
33
  #loras:
34
  #- name : "lora"
 
100
  fp8: False
101
 
102
  dataset:
 
 
 
 
 
 
 
103
  use_hdf5: True
104
  hdf5_flag: r
105
 
 
123
  sample_max_duration_batch: 300
124
  sample_shuffle: False
125
 
126
+ tasks_list: [ "tts", "stt" ]
127
+
128
  training: []
129
  validation: []
130
  noise: []