Zengwei commited on
Commit
a92d04c
·
1 Parent(s): ca25fcc

upload files

Browse files
Files changed (35) hide show
  1. README.md +1 -0
  2. data/lang_bpe_500/HLG.pt +3 -0
  3. data/lang_bpe_500/L.pt +3 -0
  4. data/lang_bpe_500/LG.pt +3 -0
  5. data/lang_bpe_500/Linv.pt +3 -0
  6. data/lang_bpe_500/bpe.model +3 -0
  7. data/lang_bpe_500/tokens.txt +502 -0
  8. data/lang_bpe_500/words.txt +0 -0
  9. data/lm/G_3_gram.pt +3 -0
  10. data/lm/G_4_gram.pt +3 -0
  11. decoding_results/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt +0 -0
  12. decoding_results/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt +0 -0
  13. decoding_results/attention-decoder-rescoring-no-ngram/log-decode-epoch-50-avg-30-use-averaged-model-2023-11-20-16-29-10 +246 -0
  14. decoding_results/attention-decoder-rescoring-no-ngram/log-decode-epoch-50-avg-30-use-averaged-model-2024-07-09-15-17-31 +247 -0
  15. decoding_results/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt +0 -0
  16. decoding_results/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt +0 -0
  17. decoding_results/attention-decoder-rescoring-no-ngram/wer-summary-test-clean-epoch-50-avg-30-use-averaged-model.txt +29 -0
  18. decoding_results/attention-decoder-rescoring-no-ngram/wer-summary-test-other-epoch-50-avg-30-use-averaged-model.txt +29 -0
  19. decoding_results/ctc-decoding/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt +0 -0
  20. decoding_results/ctc-decoding/errs-test-other-epoch-50-avg-30-use-averaged-model.txt +0 -0
  21. decoding_results/ctc-decoding/log-decode-epoch-50-avg-30-use-averaged-model-2024-07-09-15-25-09 +31 -0
  22. decoding_results/ctc-decoding/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt +0 -0
  23. decoding_results/ctc-decoding/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt +0 -0
  24. decoding_results/ctc-decoding/wer-summary-test-clean-epoch-50-avg-30-use-averaged-model.txt +2 -0
  25. decoding_results/ctc-decoding/wer-summary-test-other-epoch-50-avg-30-use-averaged-model.txt +2 -0
  26. exp/decode.sh +19 -0
  27. exp/epoch-50.pt +3 -0
  28. exp/export.sh +15 -0
  29. exp/pretrained.pt +3 -0
  30. exp/pretrained_ctc.sh +18 -0
  31. exp/train.sh +22 -0
  32. test_wavs/1089-134686-0001.wav +0 -0
  33. test_wavs/1221-135766-0001.wav +0 -0
  34. test_wavs/1221-135766-0002.wav +0 -0
  35. test_wavs/trans.txt +3 -0
README.md ADDED
@@ -0,0 +1 @@
 
 
1
+ See https://github.com/k2-fsa/icefall/pull/1389
data/lang_bpe_500/HLG.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5dbbe8b485c0cb37d11e07e8e734990f1e40a2d00fe9689d8da2e7b6fe72883
3
+ size 845007583
data/lang_bpe_500/L.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1b88996f918737fba67fbd29152018b51a537c16ce0718a2b43d5140583224e
3
+ size 19025703
data/lang_bpe_500/LG.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3bb9f021c7aad79d45dc275ba8154a430c4f660a319dcb872cd52500f25553d6
3
+ size 249852195
data/lang_bpe_500/Linv.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbc8b3687a1b8f0811a84106b3b310642566c7b1bc282a929878f9269507a2c6
3
+ size 19025703
data/lang_bpe_500/bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c53433de083c4a6ad12d034550ef22de68cec62c4f58932a7b6b8b2f1e743fa5
3
+ size 244865
data/lang_bpe_500/tokens.txt ADDED
@@ -0,0 +1,502 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <blk> 0
2
+ <sos/eos> 1
3
+ <unk> 2
4
+ S 3
5
+ ▁THE 4
6
+ ▁A 5
7
+ T 6
8
+ ▁AND 7
9
+ ED 8
10
+ ▁OF 9
11
+ ▁TO 10
12
+ E 11
13
+ D 12
14
+ N 13
15
+ ING 14
16
+ ▁IN 15
17
+ Y 16
18
+ M 17
19
+ C 18
20
+ ▁I 19
21
+ A 20
22
+ P 21
23
+ ▁HE 22
24
+ R 23
25
+ O 24
26
+ L 25
27
+ RE 26
28
+ I 27
29
+ U 28
30
+ ER 29
31
+ ▁IT 30
32
+ LY 31
33
+ ▁THAT 32
34
+ ▁WAS 33
35
+ ▁ 34
36
+ ▁S 35
37
+ AR 36
38
+ ▁BE 37
39
+ F 38
40
+ ▁C 39
41
+ IN 40
42
+ B 41
43
+ ▁FOR 42
44
+ OR 43
45
+ LE 44
46
+ ' 45
47
+ ▁HIS 46
48
+ ▁YOU 47
49
+ AL 48
50
+ ▁RE 49
51
+ V 50
52
+ ▁B 51
53
+ G 52
54
+ RI 53
55
+ ▁E 54
56
+ ▁WITH 55
57
+ ▁T 56
58
+ ▁AS 57
59
+ LL 58
60
+ ▁P 59
61
+ ▁HER 60
62
+ ST 61
63
+ ▁HAD 62
64
+ ▁SO 63
65
+ ▁F 64
66
+ W 65
67
+ CE 66
68
+ ▁IS 67
69
+ ND 68
70
+ ▁NOT 69
71
+ TH 70
72
+ ▁BUT 71
73
+ EN 72
74
+ ▁SHE 73
75
+ ▁ON 74
76
+ VE 75
77
+ ON 76
78
+ SE 77
79
+ ▁DE 78
80
+ UR 79
81
+ ▁G 80
82
+ CH 81
83
+ K 82
84
+ TER 83
85
+ ▁AT 84
86
+ IT 85
87
+ ▁ME 86
88
+ RO 87
89
+ NE 88
90
+ RA 89
91
+ ES 90
92
+ IL 91
93
+ NG 92
94
+ IC 93
95
+ ▁NO 94
96
+ ▁HIM 95
97
+ ENT 96
98
+ IR 97
99
+ ▁WE 98
100
+ H 99
101
+ ▁DO 100
102
+ ▁ALL 101
103
+ ▁HAVE 102
104
+ LO 103
105
+ ▁BY 104
106
+ ▁MY 105
107
+ ▁MO 106
108
+ ▁THIS 107
109
+ LA 108
110
+ ▁ST 109
111
+ ▁WHICH 110
112
+ ▁CON 111
113
+ ▁THEY 112
114
+ CK 113
115
+ TE 114
116
+ ▁SAID 115
117
+ ▁FROM 116
118
+ ▁GO 117
119
+ ▁WHO 118
120
+ ▁TH 119
121
+ ▁OR 120
122
+ ▁D 121
123
+ ▁W 122
124
+ VER 123
125
+ LI 124
126
+ ▁SE 125
127
+ ▁ONE 126
128
+ ▁CA 127
129
+ ▁AN 128
130
+ ▁LA 129
131
+ ▁WERE 130
132
+ EL 131
133
+ ▁HA 132
134
+ ▁MAN 133
135
+ ▁FA 134
136
+ ▁EX 135
137
+ AD 136
138
+ ▁SU 137
139
+ RY 138
140
+ ▁MI 139
141
+ AT 140
142
+ ▁BO 141
143
+ ▁WHEN 142
144
+ AN 143
145
+ THER 144
146
+ PP 145
147
+ ATION 146
148
+ ▁FI 147
149
+ ▁WOULD 148
150
+ ▁PRO 149
151
+ OW 150
152
+ ET 151
153
+ ▁O 152
154
+ ▁THERE 153
155
+ ▁HO 154
156
+ ION 155
157
+ ▁WHAT 156
158
+ ▁FE 157
159
+ ▁PA 158
160
+ US 159
161
+ MENT 160
162
+ ▁MA 161
163
+ UT 162
164
+ ▁OUT 163
165
+ ▁THEIR 164
166
+ ▁IF 165
167
+ ▁LI 166
168
+ ▁K 167
169
+ ▁WILL 168
170
+ ▁ARE 169
171
+ ID 170
172
+ ▁RO 171
173
+ DE 172
174
+ TION 173
175
+ ▁WA 174
176
+ PE 175
177
+ ▁UP 176
178
+ ▁SP 177
179
+ ▁PO 178
180
+ IGHT 179
181
+ ▁UN 180
182
+ RU 181
183
+ ▁LO 182
184
+ AS 183
185
+ OL 184
186
+ ▁LE 185
187
+ ▁BEEN 186
188
+ ▁SH 187
189
+ ▁RA 188
190
+ ▁SEE 189
191
+ KE 190
192
+ UL 191
193
+ TED 192
194
+ ▁SA 193
195
+ UN 194
196
+ UND 195
197
+ ANT 196
198
+ ▁NE 197
199
+ IS 198
200
+ ▁THEM 199
201
+ CI 200
202
+ GE 201
203
+ ▁COULD 202
204
+ ▁DIS 203
205
+ OM 204
206
+ ISH 205
207
+ HE 206
208
+ EST 207
209
+ ▁SOME 208
210
+ ENCE 209
211
+ ITY 210
212
+ IVE 211
213
+ ▁US 212
214
+ ▁MORE 213
215
+ ▁EN 214
216
+ ARD 215
217
+ ATE 216
218
+ ▁YOUR 217
219
+ ▁INTO 218
220
+ ▁KNOW 219
221
+ ▁CO 220
222
+ ANCE 221
223
+ ▁TIME 222
224
+ ▁WI 223
225
+ ▁YE 224
226
+ AGE 225
227
+ ▁NOW 226
228
+ TI 227
229
+ FF 228
230
+ ABLE 229
231
+ ▁VERY 230
232
+ ▁LIKE 231
233
+ AM 232
234
+ HI 233
235
+ Z 234
236
+ ▁OTHER 235
237
+ ▁THAN 236
238
+ ▁LITTLE 237
239
+ ▁DID 238
240
+ ▁LOOK 239
241
+ TY 240
242
+ ERS 241
243
+ ▁CAN 242
244
+ ▁CHA 243
245
+ ▁AR 244
246
+ X 245
247
+ FUL 246
248
+ UGH 247
249
+ ▁BA 248
250
+ ▁DAY 249
251
+ ▁ABOUT 250
252
+ TEN 251
253
+ IM 252
254
+ ▁ANY 253
255
+ ▁PRE 254
256
+ ▁OVER 255
257
+ IES 256
258
+ NESS 257
259
+ ME 258
260
+ BLE 259
261
+ ▁M 260
262
+ ROW 261
263
+ ▁HAS 262
264
+ ▁GREAT 263
265
+ ▁VI 264
266
+ TA 265
267
+ ▁AFTER 266
268
+ PER 267
269
+ ▁AGAIN 268
270
+ HO 269
271
+ SH 270
272
+ ▁UPON 271
273
+ ▁DI 272
274
+ ▁HAND 273
275
+ ▁COM 274
276
+ IST 275
277
+ TURE 276
278
+ ▁STA 277
279
+ ▁THEN 278
280
+ ▁SHOULD 279
281
+ ▁GA 280
282
+ OUS 281
283
+ OUR 282
284
+ ▁WELL 283
285
+ ▁ONLY 284
286
+ MAN 285
287
+ ▁GOOD 286
288
+ ▁TWO 287
289
+ ▁MAR 288
290
+ ▁SAY 289
291
+ ▁HU 290
292
+ TING 291
293
+ ▁OUR 292
294
+ RESS 293
295
+ ▁DOWN 294
296
+ IOUS 295
297
+ ▁BEFORE 296
298
+ ▁DA 297
299
+ ▁NA 298
300
+ QUI 299
301
+ ▁MADE 300
302
+ ▁EVERY 301
303
+ ▁OLD 302
304
+ ▁EVEN 303
305
+ IG 304
306
+ ▁COME 305
307
+ ▁GRA 306
308
+ ▁RI 307
309
+ ▁LONG 308
310
+ OT 309
311
+ SIDE 310
312
+ WARD 311
313
+ ▁FO 312
314
+ ▁WHERE 313
315
+ MO 314
316
+ LESS 315
317
+ ▁SC 316
318
+ ▁MUST 317
319
+ ▁NEVER 318
320
+ ▁HOW 319
321
+ ▁CAME 320
322
+ ▁SUCH 321
323
+ ▁RU 322
324
+ ▁TAKE 323
325
+ ▁WO 324
326
+ ▁CAR 325
327
+ UM 326
328
+ AK 327
329
+ ▁THINK 328
330
+ ▁MUCH 329
331
+ ▁MISTER 330
332
+ ▁MAY 331
333
+ ▁JO 332
334
+ ▁WAY 333
335
+ ▁COMP 334
336
+ ▁THOUGHT 335
337
+ ▁STO 336
338
+ ▁MEN 337
339
+ ▁BACK 338
340
+ ▁DON 339
341
+ J 340
342
+ ▁LET 341
343
+ ▁TRA 342
344
+ ▁FIRST 343
345
+ ▁JUST 344
346
+ ▁VA 345
347
+ ▁OWN 346
348
+ ▁PLA 347
349
+ ▁MAKE 348
350
+ ATED 349
351
+ ▁HIMSELF 350
352
+ ▁WENT 351
353
+ ▁PI 352
354
+ GG 353
355
+ RING 354
356
+ ▁DU 355
357
+ ▁MIGHT 356
358
+ ▁PART 357
359
+ ▁GIVE 358
360
+ ▁IMP 359
361
+ ▁BU 360
362
+ ▁PER 361
363
+ ▁PLACE 362
364
+ ▁HOUSE 363
365
+ ▁THROUGH 364
366
+ IAN 365
367
+ ▁SW 366
368
+ ▁UNDER 367
369
+ QUE 368
370
+ ▁AWAY 369
371
+ ▁LOVE 370
372
+ QUA 371
373
+ ▁LIFE 372
374
+ ▁GET 373
375
+ ▁WITHOUT 374
376
+ ▁PASS 375
377
+ ▁TURN 376
378
+ IGN 377
379
+ ▁HEAD 378
380
+ ▁MOST 379
381
+ ▁THOSE 380
382
+ ▁SHALL 381
383
+ ▁EYES 382
384
+ ▁COL 383
385
+ ▁STILL 384
386
+ ▁NIGHT 385
387
+ ▁NOTHING 386
388
+ ITION 387
389
+ HA 388
390
+ ▁TELL 389
391
+ ▁WORK 390
392
+ ▁LAST 391
393
+ ▁NEW 392
394
+ ▁FACE 393
395
+ ▁HI 394
396
+ ▁WORD 395
397
+ ▁FOUND 396
398
+ ▁COUNT 397
399
+ ▁OB 398
400
+ ▁WHILE 399
401
+ ▁SHA 400
402
+ ▁MEAN 401
403
+ ▁SAW 402
404
+ ▁PEOPLE 403
405
+ ▁FRIEND 404
406
+ ▁THREE 405
407
+ ▁ROOM 406
408
+ ▁SAME 407
409
+ ▁THOUGH 408
410
+ ▁RIGHT 409
411
+ ▁CHILD 410
412
+ ▁FATHER 411
413
+ ▁ANOTHER 412
414
+ ▁HEART 413
415
+ ▁WANT 414
416
+ ▁TOOK 415
417
+ OOK 416
418
+ ▁LIGHT 417
419
+ ▁MISSUS 418
420
+ ▁OPEN 419
421
+ ▁JU 420
422
+ ▁ASKED 421
423
+ PORT 422
424
+ ▁LEFT 423
425
+ ▁JA 424
426
+ ▁WORLD 425
427
+ ▁HOME 426
428
+ ▁WHY 427
429
+ ▁ALWAYS 428
430
+ ▁ANSWER 429
431
+ ▁SEEMED 430
432
+ ▁SOMETHING 431
433
+ ▁GIRL 432
434
+ ▁BECAUSE 433
435
+ ▁NAME 434
436
+ ▁TOLD 435
437
+ ▁NI 436
438
+ ▁HIGH 437
439
+ IZE 438
440
+ ▁WOMAN 439
441
+ ▁FOLLOW 440
442
+ ▁RETURN 441
443
+ ▁KNEW 442
444
+ ▁EACH 443
445
+ ▁KIND 444
446
+ ▁JE 445
447
+ ▁ACT 446
448
+ ▁LU 447
449
+ ▁CERTAIN 448
450
+ ▁YEARS 449
451
+ ▁QUITE 450
452
+ ▁APPEAR 451
453
+ ▁BETTER 452
454
+ ▁HALF 453
455
+ ▁PRESENT 454
456
+ ▁PRINCE 455
457
+ SHIP 456
458
+ ▁ALSO 457
459
+ ▁BEGAN 458
460
+ ▁HAVING 459
461
+ ▁ENOUGH 460
462
+ ▁PERSON 461
463
+ ▁LADY 462
464
+ ▁WHITE 463
465
+ ▁COURSE 464
466
+ ▁VOICE 465
467
+ ▁SPEAK 466
468
+ ▁POWER 467
469
+ ▁MORNING 468
470
+ ▁BETWEEN 469
471
+ ▁AMONG 470
472
+ ▁KEEP 471
473
+ ▁WALK 472
474
+ ▁MATTER 473
475
+ ▁TEA 474
476
+ ▁BELIEVE 475
477
+ ▁SMALL 476
478
+ ▁TALK 477
479
+ ▁FELT 478
480
+ ▁HORSE 479
481
+ ▁MYSELF 480
482
+ ▁SIX 481
483
+ ▁HOWEVER 482
484
+ ▁FULL 483
485
+ ▁HERSELF 484
486
+ ▁POINT 485
487
+ ▁STOOD 486
488
+ ▁HUNDRED 487
489
+ ▁ALMOST 488
490
+ ▁SINCE 489
491
+ ▁LARGE 490
492
+ ▁LEAVE 491
493
+ ▁PERHAPS 492
494
+ ▁DARK 493
495
+ ▁SUDDEN 494
496
+ ▁REPLIED 495
497
+ ▁ANYTHING 496
498
+ ▁WONDER 497
499
+ ▁UNTIL 498
500
+ Q 499
501
+ #0 500
502
+ #1 501
data/lang_bpe_500/words.txt ADDED
The diff for this file is too large to render. See raw diff
 
data/lm/G_3_gram.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:567b651ed5faa8cf83c4bef66ae330ffddb0299048754fc486293098510f3b7a
3
+ size 127948391
data/lm/G_4_gram.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c91581bc632f9c72c557ccdf726298255b9627a6ac38270c51891459b82630e9
3
+ size 3700956587
decoding_results/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/attention-decoder-rescoring-no-ngram/log-decode-epoch-50-avg-30-use-averaged-model-2023-11-20-16-29-10 ADDED
@@ -0,0 +1,246 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2023-11-20 16:29:10,461 INFO [ctc_decode.py:690] Decoding started
2
+ 2023-11-20 16:29:10,461 INFO [ctc_decode.py:696] Device: cuda:0
3
+ 2023-11-20 16:29:10,462 INFO [ctc_decode.py:697] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'ignore_id': -1, 'label_smoothing': 0.1, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.4', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '44a9d5682af9fd3ef77074777e15278ec6d390eb', 'k2-git-date': 'Wed Sep 27 11:22:55 2023', 'lhotse-version': '1.17.0.dev+git.ccfc5b2c.clean', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'zipformer-ctc-aed', 'icefall-git-sha1': '7886da9b-dirty', 'icefall-git-date': 'Tue Nov 14 20:15:43 2023', 'icefall-path': '/star-zw/workspace/zipformer/icefall_ctc_aed', 'k2-path': '/star-zw/workspace/k2/k2/k2/python/k2/__init__.py', 'lhotse-path': '/star-zw/env/k2_icefall/lib/python3.8/site-packages/lhotse-1.17.0.dev0+git.ccfc5b2c.clean-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-9-0208143539-7dbf569d4f-r7nrb', 'IP address': '10.177.13.150'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 50, 'iter': 0, 'avg': 30, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'context_size': 2, 'decoding_method': 'attention-decoder-rescoring-no-ngram', 'num_paths': 100, 'nbest_scale': 1.0, 'hlg_scale': 0.6, 'lm_dir': PosixPath('data/lm'), 'num_encoder_layers': '2,2,2,2,2,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,768,768,768,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,256,256,256,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,192,192,192,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'attention_decoder_dim': 512, 'attention_decoder_num_layers': 6, 'attention_decoder_attention_dim': 512, 'attention_decoder_num_heads': 8, 'attention_decoder_feedforward_dim': 2048, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': False, 'use_ctc': True, 'use_attention_decoder': True, 'full_libri': True, 'mini_libri': False, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 200, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram'), 'suffix': 'epoch-50-avg-30-use-averaged-model'}
4
+ 2023-11-20 16:29:10,912 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_500/Linv.pt
5
+ 2023-11-20 16:29:17,048 INFO [ctc_decode.py:780] About to create model
6
+ 2023-11-20 16:29:17,928 INFO [ctc_decode.py:847] Calculating the averaged model over epoch range from 20 (excluded) to 50
7
+ 2023-11-20 16:29:22,449 INFO [ctc_decode.py:864] Number of model parameters: 46282107
8
+ 2023-11-20 16:29:22,449 INFO [asr_datamodule.py:465] About to get test-clean cuts
9
+ 2023-11-20 16:29:22,689 INFO [asr_datamodule.py:472] About to get test-other cuts
10
+ 2023-11-20 16:29:25,562 INFO [ctc_decode.py:596] batch 0/?, cuts processed until now is 14
11
+ 2023-11-20 16:32:11,326 INFO [ctc_decode.py:596] batch 100/?, cuts processed until now is 2298
12
+ 2023-11-20 16:32:25,167 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.1159, 3.2503, 3.5065, 3.1249], device='cuda:0')
13
+ 2023-11-20 16:32:32,488 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
14
+ 2023-11-20 16:32:32,647 INFO [utils.py:641] [test-clean-attention_scale_0.01] %WER 3.09% [1626 / 52576, 151 ins, 118 del, 1357 sub ]
15
+ 2023-11-20 16:32:33,006 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
16
+ 2023-11-20 16:32:33,049 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
17
+ 2023-11-20 16:32:33,202 INFO [utils.py:641] [test-clean-attention_scale_0.05] %WER 3.02% [1588 / 52576, 148 ins, 119 del, 1321 sub ]
18
+ 2023-11-20 16:32:33,546 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
19
+ 2023-11-20 16:32:33,588 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
20
+ 2023-11-20 16:32:33,740 INFO [utils.py:641] [test-clean-attention_scale_0.08] %WER 2.99% [1571 / 52576, 150 ins, 119 del, 1302 sub ]
21
+ 2023-11-20 16:32:34,082 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
22
+ 2023-11-20 16:32:34,127 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
23
+ 2023-11-20 16:32:34,278 INFO [utils.py:641] [test-clean-attention_scale_0.1] %WER 2.98% [1565 / 52576, 151 ins, 121 del, 1293 sub ]
24
+ 2023-11-20 16:32:34,620 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
25
+ 2023-11-20 16:32:34,660 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
26
+ 2023-11-20 16:32:34,812 INFO [utils.py:641] [test-clean-attention_scale_0.3] %WER 2.78% [1461 / 52576, 145 ins, 114 del, 1202 sub ]
27
+ 2023-11-20 16:32:35,150 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
28
+ 2023-11-20 16:32:35,190 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
29
+ 2023-11-20 16:32:35,342 INFO [utils.py:641] [test-clean-attention_scale_0.5] %WER 2.69% [1413 / 52576, 146 ins, 112 del, 1155 sub ]
30
+ 2023-11-20 16:32:35,680 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
31
+ 2023-11-20 16:32:35,720 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
32
+ 2023-11-20 16:32:35,872 INFO [utils.py:641] [test-clean-attention_scale_0.6] %WER 2.67% [1402 / 52576, 147 ins, 110 del, 1145 sub ]
33
+ 2023-11-20 16:32:36,453 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
34
+ 2023-11-20 16:32:36,491 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
35
+ 2023-11-20 16:32:36,644 INFO [utils.py:641] [test-clean-attention_scale_0.7] %WER 2.64% [1389 / 52576, 145 ins, 110 del, 1134 sub ]
36
+ 2023-11-20 16:32:36,993 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
37
+ 2023-11-20 16:32:37,033 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
38
+ 2023-11-20 16:32:37,184 INFO [utils.py:641] [test-clean-attention_scale_0.9] %WER 2.62% [1376 / 52576, 148 ins, 107 del, 1121 sub ]
39
+ 2023-11-20 16:32:37,521 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
40
+ 2023-11-20 16:32:37,560 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
41
+ 2023-11-20 16:32:37,710 INFO [utils.py:641] [test-clean-attention_scale_1.0] %WER 2.60% [1368 / 52576, 147 ins, 107 del, 1114 sub ]
42
+ 2023-11-20 16:32:38,049 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
43
+ 2023-11-20 16:32:38,090 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
44
+ 2023-11-20 16:32:38,239 INFO [utils.py:641] [test-clean-attention_scale_1.1] %WER 2.59% [1360 / 52576, 146 ins, 107 del, 1107 sub ]
45
+ 2023-11-20 16:32:38,577 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
46
+ 2023-11-20 16:32:38,617 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
47
+ 2023-11-20 16:32:38,768 INFO [utils.py:641] [test-clean-attention_scale_1.2] %WER 2.58% [1357 / 52576, 146 ins, 107 del, 1104 sub ]
48
+ 2023-11-20 16:32:39,107 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
49
+ 2023-11-20 16:32:39,149 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
50
+ 2023-11-20 16:32:39,299 INFO [utils.py:641] [test-clean-attention_scale_1.3] %WER 2.58% [1354 / 52576, 147 ins, 107 del, 1100 sub ]
51
+ 2023-11-20 16:32:39,638 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
52
+ 2023-11-20 16:32:39,677 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
53
+ 2023-11-20 16:32:39,836 INFO [utils.py:641] [test-clean-attention_scale_1.5] %WER 2.54% [1337 / 52576, 143 ins, 105 del, 1089 sub ]
54
+ 2023-11-20 16:32:40,185 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
55
+ 2023-11-20 16:32:40,227 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
56
+ 2023-11-20 16:32:40,386 INFO [utils.py:641] [test-clean-attention_scale_1.7] %WER 2.53% [1332 / 52576, 142 ins, 105 del, 1085 sub ]
57
+ 2023-11-20 16:32:40,986 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
58
+ 2023-11-20 16:32:41,025 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
59
+ 2023-11-20 16:32:41,188 INFO [utils.py:641] [test-clean-attention_scale_1.9] %WER 2.52% [1327 / 52576, 139 ins, 105 del, 1083 sub ]
60
+ 2023-11-20 16:32:41,533 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
61
+ 2023-11-20 16:32:41,571 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
62
+ 2023-11-20 16:32:41,730 INFO [utils.py:641] [test-clean-attention_scale_2.0] %WER 2.52% [1326 / 52576, 139 ins, 105 del, 1082 sub ]
63
+ 2023-11-20 16:32:42,077 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
64
+ 2023-11-20 16:32:42,115 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
65
+ 2023-11-20 16:32:42,270 INFO [utils.py:641] [test-clean-attention_scale_2.1] %WER 2.51% [1321 / 52576, 139 ins, 104 del, 1078 sub ]
66
+ 2023-11-20 16:32:42,623 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
67
+ 2023-11-20 16:32:42,664 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
68
+ 2023-11-20 16:32:42,821 INFO [utils.py:641] [test-clean-attention_scale_2.2] %WER 2.50% [1316 / 52576, 139 ins, 103 del, 1074 sub ]
69
+ 2023-11-20 16:32:43,169 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
70
+ 2023-11-20 16:32:43,208 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
71
+ 2023-11-20 16:32:43,364 INFO [utils.py:641] [test-clean-attention_scale_2.3] %WER 2.50% [1314 / 52576, 139 ins, 102 del, 1073 sub ]
72
+ 2023-11-20 16:32:43,713 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
73
+ 2023-11-20 16:32:43,753 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
74
+ 2023-11-20 16:32:43,911 INFO [utils.py:641] [test-clean-attention_scale_2.5] %WER 2.50% [1314 / 52576, 138 ins, 102 del, 1074 sub ]
75
+ 2023-11-20 16:32:44,257 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
76
+ 2023-11-20 16:32:44,295 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
77
+ 2023-11-20 16:32:44,451 INFO [utils.py:641] [test-clean-attention_scale_3.0] %WER 2.48% [1306 / 52576, 136 ins, 100 del, 1070 sub ]
78
+ 2023-11-20 16:32:44,796 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
79
+ 2023-11-20 16:32:44,837 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
80
+ 2023-11-20 16:32:44,993 INFO [utils.py:641] [test-clean-attention_scale_4.0] %WER 2.47% [1301 / 52576, 134 ins, 99 del, 1068 sub ]
81
+ 2023-11-20 16:32:45,560 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
82
+ 2023-11-20 16:32:45,598 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
83
+ 2023-11-20 16:32:45,755 INFO [utils.py:641] [test-clean-attention_scale_5.0] %WER 2.47% [1297 / 52576, 134 ins, 98 del, 1065 sub ]
84
+ 2023-11-20 16:32:46,103 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
85
+ 2023-11-20 16:32:46,143 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
86
+ 2023-11-20 16:32:46,301 INFO [utils.py:641] [test-clean-attention_scale_6.0] %WER 2.46% [1295 / 52576, 134 ins, 95 del, 1066 sub ]
87
+ 2023-11-20 16:32:46,650 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
88
+ 2023-11-20 16:32:46,691 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
89
+ 2023-11-20 16:32:46,848 INFO [utils.py:641] [test-clean-attention_scale_7.0] %WER 2.47% [1297 / 52576, 135 ins, 95 del, 1067 sub ]
90
+ 2023-11-20 16:32:47,196 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
91
+ 2023-11-20 16:32:47,235 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
92
+ 2023-11-20 16:32:47,392 INFO [utils.py:641] [test-clean-attention_scale_8.0] %WER 2.46% [1295 / 52576, 136 ins, 95 del, 1064 sub ]
93
+ 2023-11-20 16:32:47,740 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
94
+ 2023-11-20 16:32:47,778 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
95
+ 2023-11-20 16:32:47,936 INFO [utils.py:641] [test-clean-attention_scale_9.0] %WER 2.47% [1296 / 52576, 136 ins, 95 del, 1065 sub ]
96
+ 2023-11-20 16:32:48,281 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
97
+ 2023-11-20 16:32:48,284 INFO [ctc_decode.py:642]
98
+ For test-clean, WER of different settings are:
99
+ attention_scale_6.0 2.46 best for test-clean
100
+ attention_scale_8.0 2.46
101
+ attention_scale_4.0 2.47
102
+ attention_scale_5.0 2.47
103
+ attention_scale_7.0 2.47
104
+ attention_scale_9.0 2.47
105
+ attention_scale_3.0 2.48
106
+ attention_scale_2.2 2.5
107
+ attention_scale_2.3 2.5
108
+ attention_scale_2.5 2.5
109
+ attention_scale_2.1 2.51
110
+ attention_scale_1.9 2.52
111
+ attention_scale_2.0 2.52
112
+ attention_scale_1.7 2.53
113
+ attention_scale_1.5 2.54
114
+ attention_scale_1.2 2.58
115
+ attention_scale_1.3 2.58
116
+ attention_scale_1.1 2.59
117
+ attention_scale_1.0 2.6
118
+ attention_scale_0.9 2.62
119
+ attention_scale_0.7 2.64
120
+ attention_scale_0.6 2.67
121
+ attention_scale_0.5 2.69
122
+ attention_scale_0.3 2.78
123
+ attention_scale_0.1 2.98
124
+ attention_scale_0.08 2.99
125
+ attention_scale_0.05 3.02
126
+ attention_scale_0.01 3.09
127
+
128
+ 2023-11-20 16:32:51,116 INFO [ctc_decode.py:596] batch 0/?, cuts processed until now is 17
129
+ 2023-11-20 16:32:57,556 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([5.4608, 5.1723, 4.6729, 4.9642], device='cuda:0')
130
+ 2023-11-20 16:35:41,583 INFO [ctc_decode.py:596] batch 100/?, cuts processed until now is 2530
131
+ 2023-11-20 16:36:01,650 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
132
+ 2023-11-20 16:36:01,812 INFO [utils.py:641] [test-other-attention_scale_0.01] %WER 7.10% [3718 / 52343, 334 ins, 285 del, 3099 sub ]
133
+ 2023-11-20 16:36:02,176 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
134
+ 2023-11-20 16:36:02,221 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
135
+ 2023-11-20 16:36:02,380 INFO [utils.py:641] [test-other-attention_scale_0.05] %WER 6.94% [3630 / 52343, 329 ins, 276 del, 3025 sub ]
136
+ 2023-11-20 16:36:02,737 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
137
+ 2023-11-20 16:36:02,780 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
138
+ 2023-11-20 16:36:02,937 INFO [utils.py:641] [test-other-attention_scale_0.08] %WER 6.85% [3583 / 52343, 325 ins, 269 del, 2989 sub ]
139
+ 2023-11-20 16:36:03,291 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
140
+ 2023-11-20 16:36:03,334 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
141
+ 2023-11-20 16:36:03,774 INFO [utils.py:641] [test-other-attention_scale_0.1] %WER 6.81% [3567 / 52343, 328 ins, 269 del, 2970 sub ]
142
+ 2023-11-20 16:36:04,126 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
143
+ 2023-11-20 16:36:04,170 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
144
+ 2023-11-20 16:36:04,326 INFO [utils.py:641] [test-other-attention_scale_0.3] %WER 6.48% [3390 / 52343, 329 ins, 255 del, 2806 sub ]
145
+ 2023-11-20 16:36:04,684 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
146
+ 2023-11-20 16:36:04,727 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
147
+ 2023-11-20 16:36:04,886 INFO [utils.py:641] [test-other-attention_scale_0.5] %WER 6.35% [3326 / 52343, 330 ins, 250 del, 2746 sub ]
148
+ 2023-11-20 16:36:05,239 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
149
+ 2023-11-20 16:36:05,282 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
150
+ 2023-11-20 16:36:05,437 INFO [utils.py:641] [test-other-attention_scale_0.6] %WER 6.31% [3302 / 52343, 329 ins, 247 del, 2726 sub ]
151
+ 2023-11-20 16:36:05,790 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
152
+ 2023-11-20 16:36:05,830 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
153
+ 2023-11-20 16:36:05,985 INFO [utils.py:641] [test-other-attention_scale_0.7] %WER 6.28% [3285 / 52343, 331 ins, 243 del, 2711 sub ]
154
+ 2023-11-20 16:36:06,342 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
155
+ 2023-11-20 16:36:06,385 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
156
+ 2023-11-20 16:36:06,542 INFO [utils.py:641] [test-other-attention_scale_0.9] %WER 6.21% [3252 / 52343, 329 ins, 244 del, 2679 sub ]
157
+ 2023-11-20 16:36:06,892 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
158
+ 2023-11-20 16:36:06,932 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
159
+ 2023-11-20 16:36:07,087 INFO [utils.py:641] [test-other-attention_scale_1.0] %WER 6.20% [3243 / 52343, 330 ins, 242 del, 2671 sub ]
160
+ 2023-11-20 16:36:07,440 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
161
+ 2023-11-20 16:36:07,480 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
162
+ 2023-11-20 16:36:07,637 INFO [utils.py:641] [test-other-attention_scale_1.1] %WER 6.18% [3237 / 52343, 332 ins, 241 del, 2664 sub ]
163
+ 2023-11-20 16:36:07,992 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
164
+ 2023-11-20 16:36:08,033 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
165
+ 2023-11-20 16:36:08,189 INFO [utils.py:641] [test-other-attention_scale_1.2] %WER 6.19% [3239 / 52343, 335 ins, 240 del, 2664 sub ]
166
+ 2023-11-20 16:36:08,773 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
167
+ 2023-11-20 16:36:08,816 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
168
+ 2023-11-20 16:36:08,973 INFO [utils.py:641] [test-other-attention_scale_1.3] %WER 6.17% [3229 / 52343, 332 ins, 240 del, 2657 sub ]
169
+ 2023-11-20 16:36:09,327 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
170
+ 2023-11-20 16:36:09,366 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
171
+ 2023-11-20 16:36:09,524 INFO [utils.py:641] [test-other-attention_scale_1.5] %WER 6.14% [3216 / 52343, 333 ins, 239 del, 2644 sub ]
172
+ 2023-11-20 16:36:09,878 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
173
+ 2023-11-20 16:36:09,921 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
174
+ 2023-11-20 16:36:10,079 INFO [utils.py:641] [test-other-attention_scale_1.7] %WER 6.12% [3201 / 52343, 332 ins, 237 del, 2632 sub ]
175
+ 2023-11-20 16:36:10,430 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
176
+ 2023-11-20 16:36:10,470 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
177
+ 2023-11-20 16:36:10,631 INFO [utils.py:641] [test-other-attention_scale_1.9] %WER 6.10% [3194 / 52343, 329 ins, 237 del, 2628 sub ]
178
+ 2023-11-20 16:36:10,981 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
179
+ 2023-11-20 16:36:11,020 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
180
+ 2023-11-20 16:36:11,175 INFO [utils.py:641] [test-other-attention_scale_2.0] %WER 6.10% [3192 / 52343, 329 ins, 236 del, 2627 sub ]
181
+ 2023-11-20 16:36:11,528 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
182
+ 2023-11-20 16:36:11,568 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
183
+ 2023-11-20 16:36:11,724 INFO [utils.py:641] [test-other-attention_scale_2.1] %WER 6.09% [3187 / 52343, 326 ins, 237 del, 2624 sub ]
184
+ 2023-11-20 16:36:12,078 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
185
+ 2023-11-20 16:36:12,119 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
186
+ 2023-11-20 16:36:12,273 INFO [utils.py:641] [test-other-attention_scale_2.2] %WER 6.09% [3186 / 52343, 326 ins, 238 del, 2622 sub ]
187
+ 2023-11-20 16:36:12,626 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
188
+ 2023-11-20 16:36:12,665 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
189
+ 2023-11-20 16:36:13,041 INFO [utils.py:641] [test-other-attention_scale_2.3] %WER 6.09% [3189 / 52343, 328 ins, 239 del, 2622 sub ]
190
+ 2023-11-20 16:36:13,390 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
191
+ 2023-11-20 16:36:13,433 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
192
+ 2023-11-20 16:36:13,590 INFO [utils.py:641] [test-other-attention_scale_2.5] %WER 6.09% [3187 / 52343, 328 ins, 239 del, 2620 sub ]
193
+ 2023-11-20 16:36:13,942 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
194
+ 2023-11-20 16:36:13,983 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
195
+ 2023-11-20 16:36:14,139 INFO [utils.py:641] [test-other-attention_scale_3.0] %WER 6.06% [3174 / 52343, 325 ins, 238 del, 2611 sub ]
196
+ 2023-11-20 16:36:14,493 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
197
+ 2023-11-20 16:36:14,532 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
198
+ 2023-11-20 16:36:14,687 INFO [utils.py:641] [test-other-attention_scale_4.0] %WER 6.05% [3166 / 52343, 324 ins, 235 del, 2607 sub ]
199
+ 2023-11-20 16:36:15,054 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
200
+ 2023-11-20 16:36:15,093 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
201
+ 2023-11-20 16:36:15,249 INFO [utils.py:641] [test-other-attention_scale_5.0] %WER 6.05% [3168 / 52343, 326 ins, 236 del, 2606 sub ]
202
+ 2023-11-20 16:36:15,601 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
203
+ 2023-11-20 16:36:15,643 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
204
+ 2023-11-20 16:36:15,799 INFO [utils.py:641] [test-other-attention_scale_6.0] %WER 6.04% [3162 / 52343, 328 ins, 236 del, 2598 sub ]
205
+ 2023-11-20 16:36:16,150 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
206
+ 2023-11-20 16:36:16,190 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
207
+ 2023-11-20 16:36:16,344 INFO [utils.py:641] [test-other-attention_scale_7.0] %WER 6.05% [3169 / 52343, 330 ins, 234 del, 2605 sub ]
208
+ 2023-11-20 16:36:16,694 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
209
+ 2023-11-20 16:36:16,735 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
210
+ 2023-11-20 16:36:16,889 INFO [utils.py:641] [test-other-attention_scale_8.0] %WER 6.06% [3173 / 52343, 330 ins, 237 del, 2606 sub ]
211
+ 2023-11-20 16:36:17,237 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
212
+ 2023-11-20 16:36:17,277 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
213
+ 2023-11-20 16:36:17,430 INFO [utils.py:641] [test-other-attention_scale_9.0] %WER 6.06% [3170 / 52343, 330 ins, 236 del, 2604 sub ]
214
+ 2023-11-20 16:36:17,998 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
215
+ 2023-11-20 16:36:18,002 INFO [ctc_decode.py:642]
216
+ For test-other, WER of different settings are:
217
+ attention_scale_6.0 6.04 best for test-other
218
+ attention_scale_4.0 6.05
219
+ attention_scale_5.0 6.05
220
+ attention_scale_7.0 6.05
221
+ attention_scale_3.0 6.06
222
+ attention_scale_8.0 6.06
223
+ attention_scale_9.0 6.06
224
+ attention_scale_2.1 6.09
225
+ attention_scale_2.2 6.09
226
+ attention_scale_2.3 6.09
227
+ attention_scale_2.5 6.09
228
+ attention_scale_1.9 6.1
229
+ attention_scale_2.0 6.1
230
+ attention_scale_1.7 6.12
231
+ attention_scale_1.5 6.14
232
+ attention_scale_1.3 6.17
233
+ attention_scale_1.1 6.18
234
+ attention_scale_1.2 6.19
235
+ attention_scale_1.0 6.2
236
+ attention_scale_0.9 6.21
237
+ attention_scale_0.7 6.28
238
+ attention_scale_0.6 6.31
239
+ attention_scale_0.5 6.35
240
+ attention_scale_0.3 6.48
241
+ attention_scale_0.1 6.81
242
+ attention_scale_0.08 6.85
243
+ attention_scale_0.05 6.94
244
+ attention_scale_0.01 7.1
245
+
246
+ 2023-11-20 16:36:18,002 INFO [ctc_decode.py:897] Done!
decoding_results/attention-decoder-rescoring-no-ngram/log-decode-epoch-50-avg-30-use-averaged-model-2024-07-09-15-17-31 ADDED
@@ -0,0 +1,247 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-07-09 15:17:31,421 INFO [ctc_decode.py:717] Decoding started
2
+ 2024-07-09 15:17:31,422 INFO [ctc_decode.py:723] Device: cuda:0
3
+ 2024-07-09 15:17:31,422 INFO [ctc_decode.py:724] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'ignore_id': -1, 'label_smoothing': 0.1, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.4', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '44a9d5682af9fd3ef77074777e15278ec6d390eb', 'k2-git-date': 'Wed Sep 27 11:22:55 2023', 'lhotse-version': '1.17.0.dev+git.ccfc5b2c.dirty', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'zipformer-ctc-aed', 'icefall-git-sha1': '4e7cdb5d-dirty', 'icefall-git-date': 'Fri Jul 5 20:13:52 2024', 'icefall-path': '/star-zw/workspace/zipformer/icefall_ctc_aed', 'k2-path': '/star-zw/workspace/k2/k2/k2/python/k2/__init__.py', 'lhotse-path': '/star-zw/workspace/lhotse/lhotse/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-1-0520143211-6b49b6c7c5-qc5gc', 'IP address': '10.177.74.203'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 50, 'iter': 0, 'avg': 30, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'context_size': 2, 'decoding_method': 'attention-decoder-rescoring-no-ngram', 'num_paths': 100, 'nbest_scale': 1.0, 'hlg_scale': 0.6, 'lm_dir': PosixPath('data/lm'), 'num_encoder_layers': '2,2,2,2,2,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,768,768,768,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,256,256,256,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,192,192,192,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'attention_decoder_dim': 512, 'attention_decoder_num_layers': 6, 'attention_decoder_attention_dim': 512, 'attention_decoder_num_heads': 8, 'attention_decoder_feedforward_dim': 2048, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': False, 'use_ctc': True, 'use_attention_decoder': True, 'full_libri': True, 'mini_libri': False, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 200, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram'), 'suffix': 'epoch-50-avg-30-use-averaged-model'}
4
+ 2024-07-09 15:17:31,713 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_500/Linv.pt
5
+ 2024-07-09 15:17:36,715 INFO [ctc_decode.py:807] About to create model
6
+ 2024-07-09 15:17:37,393 INFO [ctc_decode.py:874] Calculating the averaged model over epoch range from 20 (excluded) to 50
7
+ 2024-07-09 15:17:42,764 INFO [ctc_decode.py:891] Number of model parameters: 46282107
8
+ 2024-07-09 15:17:42,765 INFO [asr_datamodule.py:467] About to get test-clean cuts
9
+ 2024-07-09 15:17:42,887 INFO [asr_datamodule.py:474] About to get test-other cuts
10
+ 2024-07-09 15:17:45,167 INFO [ctc_decode.py:623] batch 0/?, cuts processed until now is 14
11
+ 2024-07-09 15:20:19,822 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([5.6880, 4.8939, 5.4097, 5.2731], device='cuda:0')
12
+ 2024-07-09 15:20:24,535 INFO [ctc_decode.py:623] batch 100/?, cuts processed until now is 2298
13
+ 2024-07-09 15:20:29,612 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([3.3395, 3.7145, 4.5683, 4.0622], device='cuda:0')
14
+ 2024-07-09 15:20:44,455 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
15
+ 2024-07-09 15:20:44,568 INFO [utils.py:657] [test-clean-attention_scale_0.01] %WER 3.07% [1615 / 52576, 150 ins, 120 del, 1345 sub ]
16
+ 2024-07-09 15:20:44,775 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
17
+ 2024-07-09 15:20:44,806 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
18
+ 2024-07-09 15:20:44,910 INFO [utils.py:657] [test-clean-attention_scale_0.05] %WER 3.01% [1583 / 52576, 147 ins, 121 del, 1315 sub ]
19
+ 2024-07-09 15:20:45,112 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
20
+ 2024-07-09 15:20:45,142 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
21
+ 2024-07-09 15:20:45,236 INFO [utils.py:657] [test-clean-attention_scale_0.08] %WER 2.98% [1565 / 52576, 150 ins, 119 del, 1296 sub ]
22
+ 2024-07-09 15:20:45,438 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
23
+ 2024-07-09 15:20:45,468 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
24
+ 2024-07-09 15:20:45,846 INFO [utils.py:657] [test-clean-attention_scale_0.1] %WER 2.96% [1557 / 52576, 151 ins, 120 del, 1286 sub ]
25
+ 2024-07-09 15:20:46,043 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
26
+ 2024-07-09 15:20:46,070 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
27
+ 2024-07-09 15:20:46,165 INFO [utils.py:657] [test-clean-attention_scale_0.3] %WER 2.76% [1451 / 52576, 148 ins, 112 del, 1191 sub ]
28
+ 2024-07-09 15:20:46,379 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
29
+ 2024-07-09 15:20:46,407 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
30
+ 2024-07-09 15:20:46,497 INFO [utils.py:657] [test-clean-attention_scale_0.5] %WER 2.66% [1398 / 52576, 145 ins, 110 del, 1143 sub ]
31
+ 2024-07-09 15:20:46,696 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
32
+ 2024-07-09 15:20:46,724 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
33
+ 2024-07-09 15:20:46,816 INFO [utils.py:657] [test-clean-attention_scale_0.6] %WER 2.64% [1388 / 52576, 147 ins, 107 del, 1134 sub ]
34
+ 2024-07-09 15:20:47,014 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
35
+ 2024-07-09 15:20:47,041 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
36
+ 2024-07-09 15:20:47,134 INFO [utils.py:657] [test-clean-attention_scale_0.7] %WER 2.61% [1374 / 52576, 145 ins, 106 del, 1123 sub ]
37
+ 2024-07-09 15:20:47,338 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
38
+ 2024-07-09 15:20:47,365 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
39
+ 2024-07-09 15:20:47,455 INFO [utils.py:657] [test-clean-attention_scale_0.9] %WER 2.58% [1354 / 52576, 145 ins, 104 del, 1105 sub ]
40
+ 2024-07-09 15:20:47,665 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
41
+ 2024-07-09 15:20:47,695 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
42
+ 2024-07-09 15:20:47,793 INFO [utils.py:657] [test-clean-attention_scale_1.0] %WER 2.57% [1352 / 52576, 145 ins, 104 del, 1103 sub ]
43
+ 2024-07-09 15:20:48,007 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
44
+ 2024-07-09 15:20:48,038 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
45
+ 2024-07-09 15:20:48,128 INFO [utils.py:657] [test-clean-attention_scale_1.1] %WER 2.56% [1344 / 52576, 144 ins, 105 del, 1095 sub ]
46
+ 2024-07-09 15:20:48,327 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
47
+ 2024-07-09 15:20:48,356 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
48
+ 2024-07-09 15:20:48,445 INFO [utils.py:657] [test-clean-attention_scale_1.2] %WER 2.55% [1341 / 52576, 144 ins, 105 del, 1092 sub ]
49
+ 2024-07-09 15:20:48,658 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
50
+ 2024-07-09 15:20:48,687 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
51
+ 2024-07-09 15:20:48,776 INFO [utils.py:657] [test-clean-attention_scale_1.3] %WER 2.54% [1335 / 52576, 142 ins, 105 del, 1088 sub ]
52
+ 2024-07-09 15:20:48,974 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
53
+ 2024-07-09 15:20:49,003 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
54
+ 2024-07-09 15:20:49,096 INFO [utils.py:657] [test-clean-attention_scale_1.5] %WER 2.52% [1324 / 52576, 140 ins, 102 del, 1082 sub ]
55
+ 2024-07-09 15:20:49,324 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
56
+ 2024-07-09 15:20:49,354 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
57
+ 2024-07-09 15:20:49,450 INFO [utils.py:657] [test-clean-attention_scale_1.7] %WER 2.50% [1312 / 52576, 140 ins, 102 del, 1070 sub ]
58
+ 2024-07-09 15:20:49,949 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
59
+ 2024-07-09 15:20:49,977 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
60
+ 2024-07-09 15:20:50,067 INFO [utils.py:657] [test-clean-attention_scale_1.9] %WER 2.48% [1305 / 52576, 138 ins, 100 del, 1067 sub ]
61
+ 2024-07-09 15:20:50,270 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
62
+ 2024-07-09 15:20:50,299 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
63
+ 2024-07-09 15:20:50,389 INFO [utils.py:657] [test-clean-attention_scale_2.0] %WER 2.48% [1304 / 52576, 137 ins, 100 del, 1067 sub ]
64
+ 2024-07-09 15:20:50,592 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
65
+ 2024-07-09 15:20:50,622 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
66
+ 2024-07-09 15:20:50,721 INFO [utils.py:657] [test-clean-attention_scale_2.1] %WER 2.47% [1299 / 52576, 136 ins, 100 del, 1063 sub ]
67
+ 2024-07-09 15:20:50,926 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
68
+ 2024-07-09 15:20:50,965 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
69
+ 2024-07-09 15:20:51,065 INFO [utils.py:657] [test-clean-attention_scale_2.2] %WER 2.47% [1300 / 52576, 136 ins, 100 del, 1064 sub ]
70
+ 2024-07-09 15:20:51,280 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
71
+ 2024-07-09 15:20:51,309 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
72
+ 2024-07-09 15:20:51,400 INFO [utils.py:657] [test-clean-attention_scale_2.3] %WER 2.47% [1300 / 52576, 135 ins, 99 del, 1066 sub ]
73
+ 2024-07-09 15:20:51,596 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
74
+ 2024-07-09 15:20:51,627 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
75
+ 2024-07-09 15:20:51,715 INFO [utils.py:657] [test-clean-attention_scale_2.5] %WER 2.47% [1299 / 52576, 135 ins, 99 del, 1065 sub ]
76
+ 2024-07-09 15:20:51,910 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
77
+ 2024-07-09 15:20:51,939 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
78
+ 2024-07-09 15:20:52,046 INFO [utils.py:657] [test-clean-attention_scale_3.0] %WER 2.47% [1297 / 52576, 134 ins, 99 del, 1064 sub ]
79
+ 2024-07-09 15:20:52,251 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
80
+ 2024-07-09 15:20:52,286 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
81
+ 2024-07-09 15:20:52,380 INFO [utils.py:657] [test-clean-attention_scale_4.0] %WER 2.47% [1299 / 52576, 135 ins, 100 del, 1064 sub ]
82
+ 2024-07-09 15:20:52,591 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
83
+ 2024-07-09 15:20:52,621 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
84
+ 2024-07-09 15:20:52,713 INFO [utils.py:657] [test-clean-attention_scale_5.0] %WER 2.46% [1291 / 52576, 135 ins, 98 del, 1058 sub ]
85
+ 2024-07-09 15:20:52,909 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
86
+ 2024-07-09 15:20:52,937 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
87
+ 2024-07-09 15:20:53,035 INFO [utils.py:657] [test-clean-attention_scale_6.0] %WER 2.46% [1291 / 52576, 135 ins, 97 del, 1059 sub ]
88
+ 2024-07-09 15:20:53,227 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
89
+ 2024-07-09 15:20:53,254 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
90
+ 2024-07-09 15:20:53,344 INFO [utils.py:657] [test-clean-attention_scale_7.0] %WER 2.46% [1292 / 52576, 136 ins, 97 del, 1059 sub ]
91
+ 2024-07-09 15:20:53,545 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
92
+ 2024-07-09 15:20:53,579 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
93
+ 2024-07-09 15:20:53,678 INFO [utils.py:657] [test-clean-attention_scale_8.0] %WER 2.45% [1290 / 52576, 136 ins, 97 del, 1057 sub ]
94
+ 2024-07-09 15:20:54,194 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
95
+ 2024-07-09 15:20:54,222 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
96
+ 2024-07-09 15:20:54,311 INFO [utils.py:657] [test-clean-attention_scale_9.0] %WER 2.45% [1288 / 52576, 135 ins, 97 del, 1056 sub ]
97
+ 2024-07-09 15:20:54,503 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
98
+ 2024-07-09 15:20:54,506 INFO [ctc_decode.py:669]
99
+ For test-clean, WER of different settings are:
100
+ attention_scale_8.0 2.45 best for test-clean
101
+ attention_scale_9.0 2.45
102
+ attention_scale_5.0 2.46
103
+ attention_scale_6.0 2.46
104
+ attention_scale_7.0 2.46
105
+ attention_scale_2.1 2.47
106
+ attention_scale_2.2 2.47
107
+ attention_scale_2.3 2.47
108
+ attention_scale_2.5 2.47
109
+ attention_scale_3.0 2.47
110
+ attention_scale_4.0 2.47
111
+ attention_scale_1.9 2.48
112
+ attention_scale_2.0 2.48
113
+ attention_scale_1.7 2.5
114
+ attention_scale_1.5 2.52
115
+ attention_scale_1.3 2.54
116
+ attention_scale_1.2 2.55
117
+ attention_scale_1.1 2.56
118
+ attention_scale_1.0 2.57
119
+ attention_scale_0.9 2.58
120
+ attention_scale_0.7 2.61
121
+ attention_scale_0.6 2.64
122
+ attention_scale_0.5 2.66
123
+ attention_scale_0.3 2.76
124
+ attention_scale_0.1 2.96
125
+ attention_scale_0.08 2.98
126
+ attention_scale_0.05 3.01
127
+ attention_scale_0.01 3.07
128
+
129
+ 2024-07-09 15:20:56,635 INFO [ctc_decode.py:623] batch 0/?, cuts processed until now is 17
130
+ 2024-07-09 15:22:20,800 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.7194, 3.1550, 3.8358, 3.3855], device='cuda:0')
131
+ 2024-07-09 15:23:41,174 INFO [ctc_decode.py:623] batch 100/?, cuts processed until now is 2530
132
+ 2024-07-09 15:24:00,251 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
133
+ 2024-07-09 15:24:00,350 INFO [utils.py:657] [test-other-attention_scale_0.01] %WER 7.14% [3739 / 52343, 341 ins, 286 del, 3112 sub ]
134
+ 2024-07-09 15:24:00,555 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
135
+ 2024-07-09 15:24:00,584 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
136
+ 2024-07-09 15:24:00,690 INFO [utils.py:657] [test-other-attention_scale_0.05] %WER 6.97% [3648 / 52343, 338 ins, 276 del, 3034 sub ]
137
+ 2024-07-09 15:24:00,906 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
138
+ 2024-07-09 15:24:00,942 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
139
+ 2024-07-09 15:24:01,045 INFO [utils.py:657] [test-other-attention_scale_0.08] %WER 6.86% [3591 / 52343, 333 ins, 271 del, 2987 sub ]
140
+ 2024-07-09 15:24:01,257 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
141
+ 2024-07-09 15:24:01,286 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
142
+ 2024-07-09 15:24:01,380 INFO [utils.py:657] [test-other-attention_scale_0.1] %WER 6.83% [3573 / 52343, 335 ins, 269 del, 2969 sub ]
143
+ 2024-07-09 15:24:01,583 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
144
+ 2024-07-09 15:24:01,612 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
145
+ 2024-07-09 15:24:01,706 INFO [utils.py:657] [test-other-attention_scale_0.3] %WER 6.51% [3406 / 52343, 336 ins, 254 del, 2816 sub ]
146
+ 2024-07-09 15:24:01,908 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
147
+ 2024-07-09 15:24:01,938 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
148
+ 2024-07-09 15:24:02,032 INFO [utils.py:657] [test-other-attention_scale_0.5] %WER 6.35% [3322 / 52343, 333 ins, 246 del, 2743 sub ]
149
+ 2024-07-09 15:24:02,249 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
150
+ 2024-07-09 15:24:02,285 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
151
+ 2024-07-09 15:24:02,398 INFO [utils.py:657] [test-other-attention_scale_0.6] %WER 6.31% [3302 / 52343, 334 ins, 245 del, 2723 sub ]
152
+ 2024-07-09 15:24:02,604 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
153
+ 2024-07-09 15:24:02,634 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
154
+ 2024-07-09 15:24:02,728 INFO [utils.py:657] [test-other-attention_scale_0.7] %WER 6.28% [3288 / 52343, 334 ins, 240 del, 2714 sub ]
155
+ 2024-07-09 15:24:02,932 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
156
+ 2024-07-09 15:24:02,962 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
157
+ 2024-07-09 15:24:03,056 INFO [utils.py:657] [test-other-attention_scale_0.9] %WER 6.23% [3259 / 52343, 332 ins, 243 del, 2684 sub ]
158
+ 2024-07-09 15:24:03,263 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
159
+ 2024-07-09 15:24:03,292 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
160
+ 2024-07-09 15:24:03,385 INFO [utils.py:657] [test-other-attention_scale_1.0] %WER 6.21% [3248 / 52343, 331 ins, 242 del, 2675 sub ]
161
+ 2024-07-09 15:24:03,891 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
162
+ 2024-07-09 15:24:03,919 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
163
+ 2024-07-09 15:24:04,012 INFO [utils.py:657] [test-other-attention_scale_1.1] %WER 6.19% [3239 / 52343, 330 ins, 240 del, 2669 sub ]
164
+ 2024-07-09 15:24:04,214 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
165
+ 2024-07-09 15:24:04,242 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
166
+ 2024-07-09 15:24:04,334 INFO [utils.py:657] [test-other-attention_scale_1.2] %WER 6.19% [3239 / 52343, 331 ins, 238 del, 2670 sub ]
167
+ 2024-07-09 15:24:04,542 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
168
+ 2024-07-09 15:24:04,571 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
169
+ 2024-07-09 15:24:04,662 INFO [utils.py:657] [test-other-attention_scale_1.3] %WER 6.16% [3222 / 52343, 328 ins, 239 del, 2655 sub ]
170
+ 2024-07-09 15:24:04,864 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
171
+ 2024-07-09 15:24:04,892 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
172
+ 2024-07-09 15:24:04,984 INFO [utils.py:657] [test-other-attention_scale_1.5] %WER 6.14% [3214 / 52343, 328 ins, 238 del, 2648 sub ]
173
+ 2024-07-09 15:24:05,194 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
174
+ 2024-07-09 15:24:05,224 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
175
+ 2024-07-09 15:24:05,327 INFO [utils.py:657] [test-other-attention_scale_1.7] %WER 6.12% [3205 / 52343, 329 ins, 240 del, 2636 sub ]
176
+ 2024-07-09 15:24:05,532 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
177
+ 2024-07-09 15:24:05,560 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
178
+ 2024-07-09 15:24:05,654 INFO [utils.py:657] [test-other-attention_scale_1.9] %WER 6.12% [3205 / 52343, 325 ins, 238 del, 2642 sub ]
179
+ 2024-07-09 15:24:05,864 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
180
+ 2024-07-09 15:24:05,893 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
181
+ 2024-07-09 15:24:05,985 INFO [utils.py:657] [test-other-attention_scale_2.0] %WER 6.12% [3206 / 52343, 326 ins, 238 del, 2642 sub ]
182
+ 2024-07-09 15:24:06,186 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
183
+ 2024-07-09 15:24:06,215 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
184
+ 2024-07-09 15:24:06,307 INFO [utils.py:657] [test-other-attention_scale_2.1] %WER 6.11% [3200 / 52343, 325 ins, 238 del, 2637 sub ]
185
+ 2024-07-09 15:24:06,512 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
186
+ 2024-07-09 15:24:06,541 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
187
+ 2024-07-09 15:24:06,643 INFO [utils.py:657] [test-other-attention_scale_2.2] %WER 6.11% [3200 / 52343, 325 ins, 241 del, 2634 sub ]
188
+ 2024-07-09 15:24:06,855 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
189
+ 2024-07-09 15:24:06,885 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
190
+ 2024-07-09 15:24:06,979 INFO [utils.py:657] [test-other-attention_scale_2.3] %WER 6.11% [3199 / 52343, 325 ins, 240 del, 2634 sub ]
191
+ 2024-07-09 15:24:07,185 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
192
+ 2024-07-09 15:24:07,214 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
193
+ 2024-07-09 15:24:07,308 INFO [utils.py:657] [test-other-attention_scale_2.5] %WER 6.12% [3205 / 52343, 326 ins, 242 del, 2637 sub ]
194
+ 2024-07-09 15:24:07,790 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
195
+ 2024-07-09 15:24:07,818 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
196
+ 2024-07-09 15:24:07,912 INFO [utils.py:657] [test-other-attention_scale_3.0] %WER 6.11% [3198 / 52343, 325 ins, 242 del, 2631 sub ]
197
+ 2024-07-09 15:24:08,122 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
198
+ 2024-07-09 15:24:08,152 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
199
+ 2024-07-09 15:24:08,255 INFO [utils.py:657] [test-other-attention_scale_4.0] %WER 6.10% [3194 / 52343, 325 ins, 242 del, 2627 sub ]
200
+ 2024-07-09 15:24:08,462 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
201
+ 2024-07-09 15:24:08,489 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
202
+ 2024-07-09 15:24:08,586 INFO [utils.py:657] [test-other-attention_scale_5.0] %WER 6.10% [3192 / 52343, 328 ins, 244 del, 2620 sub ]
203
+ 2024-07-09 15:24:08,787 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
204
+ 2024-07-09 15:24:08,814 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
205
+ 2024-07-09 15:24:08,906 INFO [utils.py:657] [test-other-attention_scale_6.0] %WER 6.08% [3184 / 52343, 330 ins, 242 del, 2612 sub ]
206
+ 2024-07-09 15:24:09,106 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
207
+ 2024-07-09 15:24:09,132 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
208
+ 2024-07-09 15:24:09,224 INFO [utils.py:657] [test-other-attention_scale_7.0] %WER 6.09% [3188 / 52343, 331 ins, 242 del, 2615 sub ]
209
+ 2024-07-09 15:24:09,427 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
210
+ 2024-07-09 15:24:09,456 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
211
+ 2024-07-09 15:24:09,559 INFO [utils.py:657] [test-other-attention_scale_8.0] %WER 6.10% [3191 / 52343, 332 ins, 242 del, 2617 sub ]
212
+ 2024-07-09 15:24:09,771 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
213
+ 2024-07-09 15:24:09,800 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
214
+ 2024-07-09 15:24:09,895 INFO [utils.py:657] [test-other-attention_scale_9.0] %WER 6.09% [3187 / 52343, 332 ins, 244 del, 2611 sub ]
215
+ 2024-07-09 15:24:10,111 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
216
+ 2024-07-09 15:24:10,115 INFO [ctc_decode.py:669]
217
+ For test-other, WER of different settings are:
218
+ attention_scale_6.0 6.08 best for test-other
219
+ attention_scale_7.0 6.09
220
+ attention_scale_9.0 6.09
221
+ attention_scale_4.0 6.1
222
+ attention_scale_5.0 6.1
223
+ attention_scale_8.0 6.1
224
+ attention_scale_2.1 6.11
225
+ attention_scale_2.2 6.11
226
+ attention_scale_2.3 6.11
227
+ attention_scale_3.0 6.11
228
+ attention_scale_1.7 6.12
229
+ attention_scale_1.9 6.12
230
+ attention_scale_2.0 6.12
231
+ attention_scale_2.5 6.12
232
+ attention_scale_1.5 6.14
233
+ attention_scale_1.3 6.16
234
+ attention_scale_1.1 6.19
235
+ attention_scale_1.2 6.19
236
+ attention_scale_1.0 6.21
237
+ attention_scale_0.9 6.23
238
+ attention_scale_0.7 6.28
239
+ attention_scale_0.6 6.31
240
+ attention_scale_0.5 6.35
241
+ attention_scale_0.3 6.51
242
+ attention_scale_0.1 6.83
243
+ attention_scale_0.08 6.86
244
+ attention_scale_0.05 6.97
245
+ attention_scale_0.01 7.14
246
+
247
+ 2024-07-09 15:24:10,115 INFO [ctc_decode.py:924] Done!
decoding_results/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/attention-decoder-rescoring-no-ngram/wer-summary-test-clean-epoch-50-avg-30-use-averaged-model.txt ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ settings WER
2
+ attention_scale_8.0 2.45
3
+ attention_scale_9.0 2.45
4
+ attention_scale_5.0 2.46
5
+ attention_scale_6.0 2.46
6
+ attention_scale_7.0 2.46
7
+ attention_scale_2.1 2.47
8
+ attention_scale_2.2 2.47
9
+ attention_scale_2.3 2.47
10
+ attention_scale_2.5 2.47
11
+ attention_scale_3.0 2.47
12
+ attention_scale_4.0 2.47
13
+ attention_scale_1.9 2.48
14
+ attention_scale_2.0 2.48
15
+ attention_scale_1.7 2.5
16
+ attention_scale_1.5 2.52
17
+ attention_scale_1.3 2.54
18
+ attention_scale_1.2 2.55
19
+ attention_scale_1.1 2.56
20
+ attention_scale_1.0 2.57
21
+ attention_scale_0.9 2.58
22
+ attention_scale_0.7 2.61
23
+ attention_scale_0.6 2.64
24
+ attention_scale_0.5 2.66
25
+ attention_scale_0.3 2.76
26
+ attention_scale_0.1 2.96
27
+ attention_scale_0.08 2.98
28
+ attention_scale_0.05 3.01
29
+ attention_scale_0.01 3.07
decoding_results/attention-decoder-rescoring-no-ngram/wer-summary-test-other-epoch-50-avg-30-use-averaged-model.txt ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ settings WER
2
+ attention_scale_6.0 6.08
3
+ attention_scale_7.0 6.09
4
+ attention_scale_9.0 6.09
5
+ attention_scale_4.0 6.1
6
+ attention_scale_5.0 6.1
7
+ attention_scale_8.0 6.1
8
+ attention_scale_2.1 6.11
9
+ attention_scale_2.2 6.11
10
+ attention_scale_2.3 6.11
11
+ attention_scale_3.0 6.11
12
+ attention_scale_1.7 6.12
13
+ attention_scale_1.9 6.12
14
+ attention_scale_2.0 6.12
15
+ attention_scale_2.5 6.12
16
+ attention_scale_1.5 6.14
17
+ attention_scale_1.3 6.16
18
+ attention_scale_1.1 6.19
19
+ attention_scale_1.2 6.19
20
+ attention_scale_1.0 6.21
21
+ attention_scale_0.9 6.23
22
+ attention_scale_0.7 6.28
23
+ attention_scale_0.6 6.31
24
+ attention_scale_0.5 6.35
25
+ attention_scale_0.3 6.51
26
+ attention_scale_0.1 6.83
27
+ attention_scale_0.08 6.86
28
+ attention_scale_0.05 6.97
29
+ attention_scale_0.01 7.14
decoding_results/ctc-decoding/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-test-other-epoch-50-avg-30-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/log-decode-epoch-50-avg-30-use-averaged-model-2024-07-09-15-25-09 ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-07-09 15:25:09,338 INFO [ctc_decode.py:717] Decoding started
2
+ 2024-07-09 15:25:09,338 INFO [ctc_decode.py:723] Device: cuda:0
3
+ 2024-07-09 15:25:09,338 INFO [ctc_decode.py:724] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'ignore_id': -1, 'label_smoothing': 0.1, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.4', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '44a9d5682af9fd3ef77074777e15278ec6d390eb', 'k2-git-date': 'Wed Sep 27 11:22:55 2023', 'lhotse-version': '1.17.0.dev+git.ccfc5b2c.dirty', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'zipformer-ctc-aed', 'icefall-git-sha1': '4e7cdb5d-dirty', 'icefall-git-date': 'Fri Jul 5 20:13:52 2024', 'icefall-path': '/star-zw/workspace/zipformer/icefall_ctc_aed', 'k2-path': '/star-zw/workspace/k2/k2/k2/python/k2/__init__.py', 'lhotse-path': '/star-zw/workspace/lhotse/lhotse/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-1-0520143211-6b49b6c7c5-qc5gc', 'IP address': '10.177.74.203'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 50, 'iter': 0, 'avg': 30, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'context_size': 2, 'decoding_method': 'ctc-decoding', 'num_paths': 100, 'nbest_scale': 1.0, 'hlg_scale': 0.6, 'lm_dir': PosixPath('data/lm'), 'num_encoder_layers': '2,2,2,2,2,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,768,768,768,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,256,256,256,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,192,192,192,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'attention_decoder_dim': 512, 'attention_decoder_num_layers': 6, 'attention_decoder_attention_dim': 512, 'attention_decoder_num_heads': 8, 'attention_decoder_feedforward_dim': 2048, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': False, 'use_ctc': True, 'use_attention_decoder': True, 'full_libri': True, 'mini_libri': False, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 200, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/ctc-decoding'), 'suffix': 'epoch-50-avg-30-use-averaged-model'}
4
+ 2024-07-09 15:25:09,621 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_500/Linv.pt
5
+ 2024-07-09 15:25:14,594 INFO [ctc_decode.py:807] About to create model
6
+ 2024-07-09 15:25:15,285 INFO [ctc_decode.py:874] Calculating the averaged model over epoch range from 20 (excluded) to 50
7
+ 2024-07-09 15:25:17,668 INFO [ctc_decode.py:891] Number of model parameters: 46282107
8
+ 2024-07-09 15:25:17,669 INFO [asr_datamodule.py:467] About to get test-clean cuts
9
+ 2024-07-09 15:25:17,784 INFO [asr_datamodule.py:474] About to get test-other cuts
10
+ 2024-07-09 15:25:18,386 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.7881, 4.9672, 4.9820, 3.2973], device='cuda:0')
11
+ 2024-07-09 15:25:18,682 INFO [ctc_decode.py:623] batch 0/?, cuts processed until now is 14
12
+ 2024-07-09 15:25:51,617 INFO [ctc_decode.py:623] batch 100/?, cuts processed until now is 2298
13
+ 2024-07-09 15:25:56,080 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/ctc-decoding/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt
14
+ 2024-07-09 15:25:56,172 INFO [utils.py:657] [test-clean-ctc-decoding] %WER 3.04% [1597 / 52576, 142 ins, 119 del, 1336 sub ]
15
+ 2024-07-09 15:25:56,470 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/ctc-decoding/errs-test-clean-epoch-50-avg-30-use-averaged-model.txt
16
+ 2024-07-09 15:25:56,473 INFO [ctc_decode.py:669]
17
+ For test-clean, WER of different settings are:
18
+ ctc-decoding 3.04 best for test-clean
19
+
20
+ 2024-07-09 15:25:57,159 INFO [ctc_decode.py:623] batch 0/?, cuts processed until now is 17
21
+ 2024-07-09 15:26:06,017 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([3.5442, 3.8729, 4.7403, 4.2821], device='cuda:0')
22
+ 2024-07-09 15:26:28,663 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.9135, 2.7641, 2.9316, 2.8356], device='cuda:0')
23
+ 2024-07-09 15:26:31,083 INFO [ctc_decode.py:623] batch 100/?, cuts processed until now is 2530
24
+ 2024-07-09 15:26:35,375 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/ctc-decoding/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt
25
+ 2024-07-09 15:26:35,472 INFO [utils.py:657] [test-other-ctc-decoding] %WER 7.04% [3686 / 52343, 345 ins, 270 del, 3071 sub ]
26
+ 2024-07-09 15:26:35,678 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-small/ctc-decoding/errs-test-other-epoch-50-avg-30-use-averaged-model.txt
27
+ 2024-07-09 15:26:35,681 INFO [ctc_decode.py:669]
28
+ For test-other, WER of different settings are:
29
+ ctc-decoding 7.04 best for test-other
30
+
31
+ 2024-07-09 15:26:35,681 INFO [ctc_decode.py:924] Done!
decoding_results/ctc-decoding/recogs-test-clean-epoch-50-avg-30-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-test-other-epoch-50-avg-30-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/wer-summary-test-clean-epoch-50-avg-30-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ ctc-decoding 3.04
decoding_results/ctc-decoding/wer-summary-test-other-epoch-50-avg-30-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ ctc-decoding 7.04
exp/decode.sh ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ export CUDA_VISIBLE_DEVICES="0"
2
+ for m in ctc-decoding attention-decoder-rescoring-no-ngram; do
3
+ ./zipformer/ctc_decode.py \
4
+ --epoch 50 \
5
+ --avg 30 \
6
+ --exp-dir zipformer/exp-small \
7
+ --use-ctc 1 \
8
+ --use-transducer 0 \
9
+ --use-attention-decoder 1 \
10
+ --attention-decoder-loss-scale 0.9 \
11
+ --num-encoder-layers 2,2,2,2,2,2 \
12
+ --feedforward-dim 512,768,768,768,768,768 \
13
+ --encoder-dim 192,256,256,256,256,256 \
14
+ --encoder-unmasked-dim 192,192,192,192,192,192 \
15
+ --max-duration 100 \
16
+ --causal 0 \
17
+ --num-paths 100 \
18
+ --decoding-method $m
19
+ done
exp/epoch-50.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49fcf9a2bc3bfdfe48b6c2adc5323d4c1d144fc912eb8f92b6ba8b20f38ac522
3
+ size 741381438
exp/export.sh ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ./zipformer/export.py \
2
+ --exp-dir zipformer/exp-small \
3
+ --use-ctc 1 \
4
+ --use-transducer 0 \
5
+ --use-attention-decoder 1 \
6
+ --num-encoder-layers 2,2,2,2,2,2 \
7
+ --feedforward-dim 512,768,768,768,768,768 \
8
+ --encoder-dim 192,256,256,256,256,256 \
9
+ --encoder-unmasked-dim 192,192,192,192,192,192 \
10
+ --tokens data/lang_bpe_500/tokens.txt \
11
+ --epoch 50 \
12
+ --avg 30 \
13
+ --jit 0
14
+
15
+
exp/pretrained.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5d113ceac34194e591d736b33f332ff43c1195a40f66b4eb7405cf630b2aa19
3
+ size 185495785
exp/pretrained_ctc.sh ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ export CUDA_VISIBLE_DEVICES=7
2
+ for m in ctc-decoding attention-decoder-rescoring-no-ngram; do
3
+ ./zipformer/pretrained_ctc.py \
4
+ --use-ctc 1 \
5
+ --use-transducer 0 \
6
+ --use-attention-decoder 1 \
7
+ --num-encoder-layers 2,2,2,2,2,2 \
8
+ --feedforward-dim 512,768,768,768,768,768 \
9
+ --encoder-dim 192,256,256,256,256,256 \
10
+ --encoder-unmasked-dim 192,192,192,192,192,192 \
11
+ --checkpoint ./zipformer/exp-small/pretrained.pt \
12
+ --tokens data/lang_bpe_500/tokens.txt \
13
+ --method $m \
14
+ --sample-rate 16000 \
15
+ test_wavs/1089-134686-0001.wav \
16
+ test_wavs/1221-135766-0001.wav \
17
+ test_wavs/1221-135766-0002.wav
18
+ done
exp/train.sh ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ export CUDA_VISIBLE_DEVICES="0,1"
2
+ # For non-streaming model training:
3
+ ./zipformer/train.py \
4
+ --world-size 2 \
5
+ --num-epochs 50 \
6
+ --start-epoch 1 \
7
+ --use-fp16 1 \
8
+ --exp-dir zipformer/exp-small \
9
+ --full-libri 1 \
10
+ --use-ctc 1 \
11
+ --use-transducer 0 \
12
+ --use-attention-decoder 1 \
13
+ --ctc-loss-scale 0.1 \
14
+ --attention-decoder-loss-scale 0.9 \
15
+ --num-encoder-layers 2,2,2,2,2,2 \
16
+ --feedforward-dim 512,768,768,768,768,768 \
17
+ --encoder-dim 192,256,256,256,256,256 \
18
+ --encoder-unmasked-dim 192,192,192,192,192,192 \
19
+ --max-duration 1700 \
20
+ --base-lr 0.04 \
21
+ --master-port 12345
22
+
test_wavs/1089-134686-0001.wav ADDED
Binary file (212 kB). View file
 
test_wavs/1221-135766-0001.wav ADDED
Binary file (535 kB). View file
 
test_wavs/1221-135766-0002.wav ADDED
Binary file (154 kB). View file
 
test_wavs/trans.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 1089-134686-0001 AFTER EARLY NIGHTFALL THE YELLOW LAMPS WOULD LIGHT UP HERE AND THERE THE SQUALID QUARTER OF THE BROTHELS
2
+ 1221-135766-0001 GOD AS A DIRECT CONSEQUENCE OF THE SIN WHICH MAN THUS PUNISHED HAD GIVEN HER A LOVELY CHILD WHOSE PLACE WAS ON THAT SAME DISHONOURED BOSOM TO CONNECT HER PARENT FOR EVER WITH THE RACE AND DESCENT OF MORTALS AND TO BE FINALLY A BLESSED SOUL IN HEAVEN
3
+ 1221-135766-0002 YET THESE THOUGHTS AFFECTED HESTER PRYNNE LESS WITH HOPE THAN APPREHENSION