qingzhengwang commited on
Commit
2fd82e3
·
1 Parent(s): 30d6719

Remove redundant log.

Browse files
exp_combined/lid_mms_ecapa_upcon_32_44_it0.4_shared_trainable_raw/lid_inference_test.log DELETED
@@ -1,300 +0,0 @@
1
- # python3 -m espnet2.bin.lid_inference_dist --output_dir exp_all_no_filter_raw/spk_mms_ecapa_upcon_32_44_it0.4_sharedCondProj_butUpdate_50k_lr1e-5_datasetup0.3_raw/inference/valid.accuracy.best/dev_babel_over_10s_lang_cross_train_all_no_filter_lang --dtype float32 --data_path_and_name_and_type dump/raw/dev_babel_over_10s_lang_cross_train_all_no_filter_lang/wav.scp,speech,sound --data_path_and_name_and_type dump/raw/dev_babel_over_10s_lang_cross_train_all_no_filter_lang/utt2spk,lid_labels,text --valid_batch_size 4 --lid_train_config exp_all_no_filter_raw/spk_mms_ecapa_upcon_32_44_it0.4_sharedCondProj_butUpdate_50k_lr1e-5_datasetup0.3_raw/config.yaml --lid_model_file exp_all_no_filter_raw/spk_mms_ecapa_upcon_32_44_it0.4_sharedCondProj_butUpdate_50k_lr1e-5_datasetup0.3_raw/valid.accuracy.best.pth --use_preprocessor true --fix_duration false --num_workers 32 --extract_embd false --save_every 1000 --resume true --save_embd_per_utt true --save_embd_avg_lang true --save_tsne_plot false --ngpu 1 --multiprocessing_distributed True
2
- # Started at Mon Jun 2 02:37:15 CDT 2025
3
- #
4
- /u/qwang20/miniconda3/envs/espnet2/bin/python3 /work/nvme/bbjs/qwang20/espnet/espnet2/bin/lid_inference_dist.py --output_dir exp_all_no_filter_raw/spk_mms_ecapa_upcon_32_44_it0.4_sharedCondProj_butUpdate_50k_lr1e-5_datasetup0.3_raw/inference/valid.accuracy.best/dev_babel_over_10s_lang_cross_train_all_no_filter_lang --dtype float32 --data_path_and_name_and_type dump/raw/dev_babel_over_10s_lang_cross_train_all_no_filter_lang/wav.scp,speech,sound --data_path_and_name_and_type dump/raw/dev_babel_over_10s_lang_cross_train_all_no_filter_lang/utt2spk,lid_labels,text --valid_batch_size 4 --lid_train_config exp_all_no_filter_raw/spk_mms_ecapa_upcon_32_44_it0.4_sharedCondProj_butUpdate_50k_lr1e-5_datasetup0.3_raw/config.yaml --lid_model_file exp_all_no_filter_raw/spk_mms_ecapa_upcon_32_44_it0.4_sharedCondProj_butUpdate_50k_lr1e-5_datasetup0.3_raw/valid.accuracy.best.pth --use_preprocessor true --fix_duration false --num_workers 32 --extract_embd false --save_every 1000 --resume true --save_embd_per_utt true --save_embd_avg_lang true --save_tsne_plot false --ngpu 1 --multiprocessing_distributed True
5
- [gpue04] 2025-06-02 02:37:35,038 (abs_task:2406) INFO: config file: exp_all_no_filter_raw/spk_mms_ecapa_upcon_32_44_it0.4_sharedCondProj_butUpdate_50k_lr1e-5_datasetup0.3_raw/config.yaml
6
- /work/nvme/bbjs/qwang20/s3prl/s3prl/upstream/byol_s/byol_a/common.py:20: UserWarning: torchaudio._backend.set_audio_backend has been deprecated. With dispatcher enabled, this function is no-op. You can remove the function call.
7
- torchaudio.set_audio_backend("sox_io")
8
- /work/nvme/bbjs/qwang20/espnet/espnet2/tasks/abs_task.py:2429: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.
9
- torch.load(model_file, map_location=device),
10
- [gpue04] 2025-06-02 02:37:46,607 (lid_inference_dist:86) INFO: Model structure:
11
- ESPnetLIDUpstreamConditionModel(
12
- (frontend): S3prlFrontendCondition(
13
- (upstream): S3PRLUpstreamCondition(
14
- (upstream): UpstreamExpertCondition(
15
- (model): Wav2Vec2ModelCondition(
16
- (feature_extractor): Wav2Vec2FeatureEncoder(
17
- (conv_layers): ModuleList(
18
- (0): Wav2Vec2LayerNormConvLayer(
19
- (conv): Conv1d(1, 512, kernel_size=(10,), stride=(5,))
20
- (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
21
- (activation): GELUActivation()
22
- )
23
- (1-4): 4 x Wav2Vec2LayerNormConvLayer(
24
- (conv): Conv1d(512, 512, kernel_size=(3,), stride=(2,))
25
- (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
26
- (activation): GELUActivation()
27
- )
28
- (5-6): 2 x Wav2Vec2LayerNormConvLayer(
29
- (conv): Conv1d(512, 512, kernel_size=(2,), stride=(2,))
30
- (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
31
- (activation): GELUActivation()
32
- )
33
- )
34
- )
35
- (feature_projection): Wav2Vec2FeatureProjection(
36
- (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
37
- (projection): Linear(in_features=512, out_features=1280, bias=True)
38
- (dropout): Dropout(p=0.1, inplace=False)
39
- )
40
- (encoder): Wav2Vec2EncoderCondition(
41
- (pos_conv_embed): Wav2Vec2PositionalConvEmbedding(
42
- (conv): ParametrizedConv1d(
43
- 1280, 1280, kernel_size=(128,), stride=(1,), padding=(64,), groups=16
44
- (parametrizations): ModuleDict(
45
- (weight): ParametrizationList(
46
- (0): _WeightNorm()
47
- )
48
- )
49
- )
50
- (padding): Wav2Vec2SamePadLayer()
51
- (activation): GELUActivation()
52
- )
53
- (layer_norm): LayerNorm((1280,), eps=1e-05, elementwise_affine=True)
54
- (dropout): Dropout(p=0.1, inplace=False)
55
- (layers): ModuleList(
56
- (0-47): 48 x Wav2Vec2EncoderLayerStableLayerNorm(
57
- (attention): Wav2Vec2SdpaAttention(
58
- (k_proj): Linear(in_features=1280, out_features=1280, bias=True)
59
- (v_proj): Linear(in_features=1280, out_features=1280, bias=True)
60
- (q_proj): Linear(in_features=1280, out_features=1280, bias=True)
61
- (out_proj): Linear(in_features=1280, out_features=1280, bias=True)
62
- )
63
- (dropout): Dropout(p=0.1, inplace=False)
64
- (layer_norm): LayerNorm((1280,), eps=1e-05, elementwise_affine=True)
65
- (feed_forward): Wav2Vec2FeedForward(
66
- (intermediate_dropout): Dropout(p=0.0, inplace=False)
67
- (intermediate_dense): Linear(in_features=1280, out_features=5120, bias=True)
68
- (intermediate_act_fn): GELUActivation()
69
- (output_dense): Linear(in_features=5120, out_features=1280, bias=True)
70
- (output_dropout): Dropout(p=0.1, inplace=False)
71
- )
72
- (final_layer_norm): LayerNorm((1280,), eps=1e-05, elementwise_affine=True)
73
- )
74
- )
75
- (ecapa_encoder): ModuleDict(
76
- (32): IdentityEncoder()
77
- (36): IdentityEncoder()
78
- (40): IdentityEncoder()
79
- (44): IdentityEncoder()
80
- )
81
- (pooling): ModuleDict(
82
- (32): ChnAttnStatPooling(
83
- (attention): Sequential(
84
- (0): Conv1d(3840, 128, kernel_size=(1,), stride=(1,))
85
- (1): ReLU()
86
- (2): BatchNorm1d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
87
- (3): Conv1d(128, 1280, kernel_size=(1,), stride=(1,))
88
- )
89
- (softmax): Softmax(dim=2)
90
- )
91
- (36): ChnAttnStatPooling(
92
- (attention): Sequential(
93
- (0): Conv1d(3840, 128, kernel_size=(1,), stride=(1,))
94
- (1): ReLU()
95
- (2): BatchNorm1d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
96
- (3): Conv1d(128, 1280, kernel_size=(1,), stride=(1,))
97
- )
98
- (softmax): Softmax(dim=2)
99
- )
100
- (40): ChnAttnStatPooling(
101
- (attention): Sequential(
102
- (0): Conv1d(3840, 128, kernel_size=(1,), stride=(1,))
103
- (1): ReLU()
104
- (2): BatchNorm1d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
105
- (3): Conv1d(128, 1280, kernel_size=(1,), stride=(1,))
106
- )
107
- (softmax): Softmax(dim=2)
108
- )
109
- (44): ChnAttnStatPooling(
110
- (attention): Sequential(
111
- (0): Conv1d(3840, 128, kernel_size=(1,), stride=(1,))
112
- (1): ReLU()
113
- (2): BatchNorm1d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
114
- (3): Conv1d(128, 1280, kernel_size=(1,), stride=(1,))
115
- )
116
- (softmax): Softmax(dim=2)
117
- )
118
- )
119
- (projector): ModuleDict(
120
- (32): RawNet3Projector(
121
- (bn): BatchNorm1d(2560, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
122
- (fc): Linear(in_features=2560, out_features=192, bias=True)
123
- )
124
- (36): RawNet3Projector(
125
- (bn): BatchNorm1d(2560, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
126
- (fc): Linear(in_features=2560, out_features=192, bias=True)
127
- )
128
- (40): RawNet3Projector(
129
- (bn): BatchNorm1d(2560, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
130
- (fc): Linear(in_features=2560, out_features=192, bias=True)
131
- )
132
- (44): RawNet3Projector(
133
- (bn): BatchNorm1d(2560, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
134
- (fc): Linear(in_features=2560, out_features=192, bias=True)
135
- )
136
- )
137
- (lang2vec_head): ModuleDict(
138
- (32): Sequential(
139
- (0): Linear(in_features=192, out_features=299, bias=True)
140
- )
141
- (36): Sequential(
142
- (0): Linear(in_features=192, out_features=299, bias=True)
143
- )
144
- (40): Sequential(
145
- (0): Linear(in_features=192, out_features=299, bias=True)
146
- )
147
- (44): Sequential(
148
- (0): Linear(in_features=192, out_features=299, bias=True)
149
- )
150
- )
151
- (aamsoftmax_weight): ParameterDict()
152
- (lang2vec_conditioning_projs): Linear(in_features=299, out_features=1280, bias=True)
153
- (aamsoftmax_loss): AAMSoftmaxSCTopKLang2Vec(
154
- (ce): CrossEntropyLoss()
155
- (lang2vec_head): Sequential(
156
- (0): Linear(in_features=192, out_features=299, bias=True)
157
- )
158
- (lang2vec_loss): MSELoss()
159
- )
160
- )
161
- )
162
- )
163
- )
164
- (featurizer): Featurizer()
165
- )
166
- (normalize): UtteranceMVN(norm_means=True, norm_vars=False)
167
- (encoder): EcapaTdnnEncoder(
168
- (conv): Conv1d(1280, 512, kernel_size=(5,), stride=(1,), padding=(2,))
169
- (relu): ReLU()
170
- (bn): BatchNorm1d(512, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
171
- (layer1): EcapaBlock(
172
- (conv1): Conv1d(512, 512, kernel_size=(1,), stride=(1,))
173
- (bn1): BatchNorm1d(512, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
174
- (convs): ModuleList(
175
- (0-6): 7 x Conv1d(64, 64, kernel_size=(3,), stride=(1,), padding=(2,), dilation=(2,))
176
- )
177
- (bns): ModuleList(
178
- (0-6): 7 x BatchNorm1d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
179
- )
180
- (conv3): Conv1d(512, 512, kernel_size=(1,), stride=(1,))
181
- (bn3): BatchNorm1d(512, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
182
- (relu): ReLU()
183
- (se): SEModule(
184
- (se): Sequential(
185
- (0): AdaptiveAvgPool1d(output_size=1)
186
- (1): Conv1d(512, 128, kernel_size=(1,), stride=(1,))
187
- (2): ReLU()
188
- (3): BatchNorm1d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
189
- (4): Conv1d(128, 512, kernel_size=(1,), stride=(1,))
190
- (5): Sigmoid()
191
- )
192
- )
193
- )
194
- (layer2): EcapaBlock(
195
- (conv1): Conv1d(512, 512, kernel_size=(1,), stride=(1,))
196
- (bn1): BatchNorm1d(512, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
197
- (convs): ModuleList(
198
- (0-6): 7 x Conv1d(64, 64, kernel_size=(3,), stride=(1,), padding=(3,), dilation=(3,))
199
- )
200
- (bns): ModuleList(
201
- (0-6): 7 x BatchNorm1d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
202
- )
203
- (conv3): Conv1d(512, 512, kernel_size=(1,), stride=(1,))
204
- (bn3): BatchNorm1d(512, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
205
- (relu): ReLU()
206
- (se): SEModule(
207
- (se): Sequential(
208
- (0): AdaptiveAvgPool1d(output_size=1)
209
- (1): Conv1d(512, 128, kernel_size=(1,), stride=(1,))
210
- (2): ReLU()
211
- (3): BatchNorm1d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
212
- (4): Conv1d(128, 512, kernel_size=(1,), stride=(1,))
213
- (5): Sigmoid()
214
- )
215
- )
216
- )
217
- (layer3): EcapaBlock(
218
- (conv1): Conv1d(512, 512, kernel_size=(1,), stride=(1,))
219
- (bn1): BatchNorm1d(512, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
220
- (convs): ModuleList(
221
- (0-6): 7 x Conv1d(64, 64, kernel_size=(3,), stride=(1,), padding=(4,), dilation=(4,))
222
- )
223
- (bns): ModuleList(
224
- (0-6): 7 x BatchNorm1d(64, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
225
- )
226
- (conv3): Conv1d(512, 512, kernel_size=(1,), stride=(1,))
227
- (bn3): BatchNorm1d(512, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
228
- (relu): ReLU()
229
- (se): SEModule(
230
- (se): Sequential(
231
- (0): AdaptiveAvgPool1d(output_size=1)
232
- (1): Conv1d(512, 128, kernel_size=(1,), stride=(1,))
233
- (2): ReLU()
234
- (3): BatchNorm1d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
235
- (4): Conv1d(128, 512, kernel_size=(1,), stride=(1,))
236
- (5): Sigmoid()
237
- )
238
- )
239
- )
240
- (layer4): Conv1d(1536, 1536, kernel_size=(1,), stride=(1,))
241
- (mp3): MaxPool1d(kernel_size=3, stride=3, padding=0, dilation=1, ceil_mode=False)
242
- )
243
- (pooling): ChnAttnStatPooling(
244
- (attention): Sequential(
245
- (0): Conv1d(4608, 128, kernel_size=(1,), stride=(1,))
246
- (1): ReLU()
247
- (2): BatchNorm1d(128, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
248
- (3): Conv1d(128, 1536, kernel_size=(1,), stride=(1,))
249
- )
250
- (softmax): Softmax(dim=2)
251
- )
252
- (projector): RawNet3Projector(
253
- (bn): BatchNorm1d(3072, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
254
- (fc): Linear(in_features=3072, out_features=192, bias=True)
255
- )
256
- (loss): AAMSoftmaxSCTopKLang2Vec(
257
- (ce): CrossEntropyLoss()
258
- (lang2vec_head): Sequential(
259
- (0): Linear(in_features=192, out_features=299, bias=True)
260
- )
261
- (lang2vec_loss): MSELoss()
262
- )
263
- )
264
-
265
- Model summary:
266
- Class Name: ESPnetLIDUpstreamConditionModel
267
- Total Number of model parameters: 977.14 M
268
- Number of trainable parameters: 977.14 M (100.0%)
269
- Size: 3.91 GB
270
- Type: torch.float32
271
- /u/qwang20/miniconda3/envs/espnet2/lib/python3.11/site-packages/torch/utils/data/dataloader.py:557: UserWarning: This DataLoader will create 32 worker processes in total. Our suggested max number of worker in current system is 16, which is smaller than what this DataLoader is going to create. Please be aware that excessive worker creation might get DataLoader running slow or even freeze, lower the worker number to avoid potential slowness/freeze if necessary.
272
- warnings.warn(_create_warning_msg(
273
- /work/nvme/bbjs/qwang20/espnet/espnet2/train/reporter.py:321: UserWarning: The stats of the previous epoch=-1doesn't exist.
274
- warnings.warn(
275
- [gpue04] 2025-06-02 02:37:47,156 (lid_trainer:102) INFO: [Rank 0] Resume: 0 utterances found in exp_all_no_filter_raw/spk_mms_ecapa_upcon_32_44_it0.4_sharedCondProj_butUpdate_50k_lr1e-5_datasetup0.3_raw/inference/valid.accuracy.best/dev_babel_over_10s_lang_cross_train_all_no_filter_lang/lids0
276
- [gpue04] 2025-06-02 02:38:41,828 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 0
277
- [gpue04] 2025-06-02 02:39:27,483 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 1
278
- [gpue04] 2025-06-02 02:40:15,909 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 2
279
- [gpue04] 2025-06-02 02:41:08,571 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 3
280
- [gpue04] 2025-06-02 02:41:56,182 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 4
281
- [gpue04] 2025-06-02 02:42:40,736 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 5
282
- [gpue04] 2025-06-02 02:43:27,814 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 6
283
- [gpue04] 2025-06-02 02:44:10,740 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 7
284
- [gpue04] 2025-06-02 02:44:52,065 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 8
285
- [gpue04] 2025-06-02 02:45:40,635 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 9
286
- [gpue04] 2025-06-02 02:46:28,394 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 10
287
- [gpue04] 2025-06-02 02:47:09,502 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 11
288
- [gpue04] 2025-06-02 02:47:59,978 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 12
289
- [gpue04] 2025-06-02 02:48:52,866 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 13
290
- [gpue04] 2025-06-02 02:49:41,279 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 14
291
- [gpue04] 2025-06-02 02:50:32,817 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 15
292
- [gpue04] 2025-06-02 02:51:20,444 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 16
293
- [gpue04] 2025-06-02 02:52:09,714 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 17
294
- [gpue04] 2025-06-02 02:52:55,108 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 18
295
- [gpue04] 2025-06-02 02:53:50,212 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 19
296
- [gpue04] 2025-06-02 02:54:31,533 (lid_trainer:207) INFO: [Rank 0] Saved 1000 utts at step 20
297
- [gpue04] 2025-06-02 02:55:19,223 (lid_inference_dist:200) INFO: args.save_embd_per_utt: True
298
- [gpue04] 2025-06-02 02:55:19,224 (lid_inference_dist:215) INFO: args.save_tsne_plot: False
299
- # Accounting: time=1085 threads=1
300
- # Ended (code 0) at Mon Jun 2 02:55:20 CDT 2025, elapsed time 1085 seconds