ChenChenyu commited on
Commit
3a88be1
·
verified ·
1 Parent(s): f005f4c

Add files using upload-large-folder tool

Browse files
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ 2024-10-25/18-23-57/train_ds.log filter=lfs diff=lfs merge=lfs -text
37
+ model.pt.best filter=lfs diff=lfs merge=lfs -text
.ipynb_checkpoints/config-checkpoint.yaml ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ encoder: SenseVoiceEncoderSmall
2
+ encoder_conf:
3
+ output_size: 512
4
+ attention_heads: 4
5
+ linear_units: 2048
6
+ num_blocks: 50
7
+ tp_blocks: 20
8
+ dropout_rate: 0.1
9
+ positional_dropout_rate: 0.1
10
+ attention_dropout_rate: 0.1
11
+ input_layer: pe
12
+ pos_enc_class: SinusoidalPositionEncoder
13
+ normalize_before: true
14
+ kernel_size: 11
15
+ sanm_shfit: 0
16
+ selfattention_layer_type: sanm
17
+ model: SenseVoiceSmall
18
+ model_conf:
19
+ length_normalized_loss: true
20
+ sos: 1
21
+ eos: 2
22
+ ignore_id: -1
23
+ tokenizer: SentencepiecesTokenizer
24
+ tokenizer_conf:
25
+ bpemodel: /root/.cache/modelscope/hub/iic/SenseVoiceSmall/chn_jpn_yue_eng_ko_spectok.bpe.model
26
+ unk_symbol: <unk>
27
+ split_with_space: true
28
+ frontend: WavFrontend
29
+ frontend_conf:
30
+ fs: 16000
31
+ window: hamming
32
+ n_mels: 80
33
+ frame_length: 25
34
+ frame_shift: 10
35
+ lfr_m: 7
36
+ lfr_n: 6
37
+ cmvn_file: /root/.cache/modelscope/hub/iic/SenseVoiceSmall/am.mvn
38
+ dataset: SenseVoiceCTCDataset
39
+ dataset_conf:
40
+ index_ds: IndexDSJsonl
41
+ batch_sampler: BatchSampler
42
+ data_split_num: 1
43
+ batch_type: token
44
+ batch_size: 25000
45
+ max_token_length: 2000
46
+ min_token_length: 60
47
+ max_source_length: 2000
48
+ min_source_length: 60
49
+ max_target_length: 200
50
+ min_target_length: 0
51
+ shuffle: true
52
+ num_workers: 4
53
+ sos: 1
54
+ eos: 2
55
+ IndexDSJsonl: IndexDSJsonl
56
+ retry: 20
57
+ sort_size: 1024
58
+ train_conf:
59
+ accum_grad: 1
60
+ grad_clip: 5
61
+ max_epoch: 100
62
+ keep_nbest_models: 10
63
+ avg_nbest_model: 10
64
+ log_interval: 1
65
+ resume: true
66
+ validate_interval: 2000
67
+ save_checkpoint_interval: 2000
68
+ use_deepspeed: false
69
+ deepspeed_config: /root/autodl-tmp/SenseVoice/deepspeed_conf/ds_stage1.json
70
+ optim: adamw
71
+ optim_conf:
72
+ lr: 0.0002
73
+ scheduler: warmuplr
74
+ scheduler_conf:
75
+ warmup_steps: 25000
76
+ specaug: SpecAugLFR
77
+ specaug_conf:
78
+ apply_time_warp: false
79
+ time_warp_window: 5
80
+ time_warp_mode: bicubic
81
+ apply_freq_mask: true
82
+ freq_mask_width_range:
83
+ - 0
84
+ - 30
85
+ lfr_rate: 6
86
+ num_freq_mask: 1
87
+ apply_time_mask: true
88
+ time_mask_width_range:
89
+ - 0
90
+ - 12
91
+ num_time_mask: 1
92
+ init_param: /root/.cache/modelscope/hub/iic/SenseVoiceSmall/model.pt
93
+ config: /root/.cache/modelscope/hub/iic/SenseVoiceSmall/config.yaml
94
+ is_training: true
95
+ trust_remote_code: true
96
+ train_data_set_list: /root/autodl-tmp/train/train.jsonl
97
+ valid_data_set_list: /root/autodl-tmp/val/val.jsonl
98
+ output_dir: ./outputs
99
+ model_path: /root/.cache/modelscope/hub/iic/SenseVoiceSmall
100
+ device: cpu
1best_recog/text ADDED
The diff for this file is too large to render. See raw diff
 
2024-10-25/18-23-57/.hydra/config.yaml ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model: iic/SenseVoiceSmall
2
+ trust_remote_code: true
3
+ train_data_set_list: /root/autodl-tmp/train/train.jsonl
4
+ valid_data_set_list: /root/autodl-tmp/val/val.jsonl
5
+ dataset_conf:
6
+ data_split_num: 1
7
+ batch_sampler: BatchSampler
8
+ batch_size: 25000
9
+ sort_size: 1024
10
+ batch_type: token
11
+ num_workers: 4
12
+ train_conf:
13
+ max_epoch: 100
14
+ log_interval: 1
15
+ resume: true
16
+ validate_interval: 2000
17
+ save_checkpoint_interval: 2000
18
+ keep_nbest_models: 10
19
+ avg_nbest_model: 10
20
+ use_deepspeed: false
21
+ deepspeed_config: /root/autodl-tmp/SenseVoice/deepspeed_conf/ds_stage1.json
22
+ optim_conf:
23
+ lr: 0.0002
24
+ output_dir: ./outputs
2024-10-25/18-23-57/.hydra/hydra.yaml ADDED
@@ -0,0 +1,172 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - ++model=iic/SenseVoiceSmall
116
+ - ++trust_remote_code=true
117
+ - ++train_data_set_list=/root/autodl-tmp/train/train.jsonl
118
+ - ++valid_data_set_list=/root/autodl-tmp/val/val.jsonl
119
+ - ++dataset_conf.data_split_num=1
120
+ - ++dataset_conf.batch_sampler=BatchSampler
121
+ - ++dataset_conf.batch_size=25000
122
+ - ++dataset_conf.sort_size=1024
123
+ - ++dataset_conf.batch_type=token
124
+ - ++dataset_conf.num_workers=4
125
+ - ++train_conf.max_epoch=100
126
+ - ++train_conf.log_interval=1
127
+ - ++train_conf.resume=true
128
+ - ++train_conf.validate_interval=2000
129
+ - ++train_conf.save_checkpoint_interval=2000
130
+ - ++train_conf.keep_nbest_models=10
131
+ - ++train_conf.avg_nbest_model=10
132
+ - ++train_conf.use_deepspeed=false
133
+ - ++train_conf.deepspeed_config=/root/autodl-tmp/SenseVoice/deepspeed_conf/ds_stage1.json
134
+ - ++optim_conf.lr=0.0002
135
+ - ++output_dir=./outputs
136
+ job:
137
+ name: train_ds
138
+ chdir: null
139
+ override_dirname: ++dataset_conf.batch_sampler=BatchSampler,++dataset_conf.batch_size=25000,++dataset_conf.batch_type=token,++dataset_conf.data_split_num=1,++dataset_conf.num_workers=4,++dataset_conf.sort_size=1024,++model=iic/SenseVoiceSmall,++optim_conf.lr=0.0002,++output_dir=./outputs,++train_conf.avg_nbest_model=10,++train_conf.deepspeed_config=/root/autodl-tmp/SenseVoice/deepspeed_conf/ds_stage1.json,++train_conf.keep_nbest_models=10,++train_conf.log_interval=1,++train_conf.max_epoch=100,++train_conf.resume=true,++train_conf.save_checkpoint_interval=2000,++train_conf.use_deepspeed=false,++train_conf.validate_interval=2000,++train_data_set_list=/root/autodl-tmp/train/train.jsonl,++trust_remote_code=true,++valid_data_set_list=/root/autodl-tmp/val/val.jsonl
140
+ id: ???
141
+ num: ???
142
+ config_name: null
143
+ env_set: {}
144
+ env_copy: []
145
+ config:
146
+ override_dirname:
147
+ kv_sep: '='
148
+ item_sep: ','
149
+ exclude_keys: []
150
+ runtime:
151
+ version: 1.3.2
152
+ version_base: '1.3'
153
+ cwd: /root/autodl-tmp/SenseVoice
154
+ config_sources:
155
+ - path: hydra.conf
156
+ schema: pkg
157
+ provider: hydra
158
+ - path: ''
159
+ schema: structured
160
+ provider: schema
161
+ output_dir: /root/autodl-tmp/SenseVoice/outputs/2024-10-25/18-23-57
162
+ choices:
163
+ hydra/env: default
164
+ hydra/callbacks: null
165
+ hydra/job_logging: default
166
+ hydra/hydra_logging: default
167
+ hydra/hydra_help: default
168
+ hydra/help: default
169
+ hydra/sweeper: basic
170
+ hydra/launcher: basic
171
+ hydra/output: default
172
+ verbose: false
2024-10-25/18-23-57/.hydra/overrides.yaml ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - ++model=iic/SenseVoiceSmall
2
+ - ++trust_remote_code=true
3
+ - ++train_data_set_list=/root/autodl-tmp/train/train.jsonl
4
+ - ++valid_data_set_list=/root/autodl-tmp/val/val.jsonl
5
+ - ++dataset_conf.data_split_num=1
6
+ - ++dataset_conf.batch_sampler=BatchSampler
7
+ - ++dataset_conf.batch_size=25000
8
+ - ++dataset_conf.sort_size=1024
9
+ - ++dataset_conf.batch_type=token
10
+ - ++dataset_conf.num_workers=4
11
+ - ++train_conf.max_epoch=100
12
+ - ++train_conf.log_interval=1
13
+ - ++train_conf.resume=true
14
+ - ++train_conf.validate_interval=2000
15
+ - ++train_conf.save_checkpoint_interval=2000
16
+ - ++train_conf.keep_nbest_models=10
17
+ - ++train_conf.avg_nbest_model=10
18
+ - ++train_conf.use_deepspeed=false
19
+ - ++train_conf.deepspeed_config=/root/autodl-tmp/SenseVoice/deepspeed_conf/ds_stage1.json
20
+ - ++optim_conf.lr=0.0002
21
+ - ++output_dir=./outputs
2024-10-25/18-23-57/train_ds.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:078bf9a7f021d04db46a4ab17a2f56246473f90ea4a12da0afd619f81f0dd4f0
3
+ size 30254298
config.yaml ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ encoder: SenseVoiceEncoderSmall
2
+ encoder_conf:
3
+ output_size: 512
4
+ attention_heads: 4
5
+ linear_units: 2048
6
+ num_blocks: 50
7
+ tp_blocks: 20
8
+ dropout_rate: 0.1
9
+ positional_dropout_rate: 0.1
10
+ attention_dropout_rate: 0.1
11
+ input_layer: pe
12
+ pos_enc_class: SinusoidalPositionEncoder
13
+ normalize_before: true
14
+ kernel_size: 11
15
+ sanm_shfit: 0
16
+ selfattention_layer_type: sanm
17
+ model: SenseVoiceSmall
18
+ model_conf:
19
+ length_normalized_loss: true
20
+ sos: 1
21
+ eos: 2
22
+ ignore_id: -1
23
+ tokenizer: SentencepiecesTokenizer
24
+ tokenizer_conf:
25
+ bpemodel: /root/.cache/modelscope/hub/iic/SenseVoiceSmall/chn_jpn_yue_eng_ko_spectok.bpe.model
26
+ unk_symbol: <unk>
27
+ split_with_space: true
28
+ frontend: WavFrontend
29
+ frontend_conf:
30
+ fs: 16000
31
+ window: hamming
32
+ n_mels: 80
33
+ frame_length: 25
34
+ frame_shift: 10
35
+ lfr_m: 7
36
+ lfr_n: 6
37
+ cmvn_file: /root/.cache/modelscope/hub/iic/SenseVoiceSmall/am.mvn
38
+ dataset: SenseVoiceCTCDataset
39
+ dataset_conf:
40
+ index_ds: IndexDSJsonl
41
+ batch_sampler: BatchSampler
42
+ data_split_num: 1
43
+ batch_type: token
44
+ batch_size: 25000
45
+ max_token_length: 2000
46
+ min_token_length: 60
47
+ max_source_length: 2000
48
+ min_source_length: 60
49
+ max_target_length: 200
50
+ min_target_length: 0
51
+ shuffle: true
52
+ num_workers: 4
53
+ sos: 1
54
+ eos: 2
55
+ IndexDSJsonl: IndexDSJsonl
56
+ retry: 20
57
+ sort_size: 1024
58
+ train_conf:
59
+ accum_grad: 1
60
+ grad_clip: 5
61
+ max_epoch: 100
62
+ keep_nbest_models: 10
63
+ avg_nbest_model: 10
64
+ log_interval: 1
65
+ resume: true
66
+ validate_interval: 2000
67
+ save_checkpoint_interval: 2000
68
+ use_deepspeed: false
69
+ deepspeed_config: /root/autodl-tmp/SenseVoice/deepspeed_conf/ds_stage1.json
70
+ optim: adamw
71
+ optim_conf:
72
+ lr: 0.0002
73
+ scheduler: warmuplr
74
+ scheduler_conf:
75
+ warmup_steps: 25000
76
+ specaug: SpecAugLFR
77
+ specaug_conf:
78
+ apply_time_warp: false
79
+ time_warp_window: 5
80
+ time_warp_mode: bicubic
81
+ apply_freq_mask: true
82
+ freq_mask_width_range:
83
+ - 0
84
+ - 30
85
+ lfr_rate: 6
86
+ num_freq_mask: 1
87
+ apply_time_mask: true
88
+ time_mask_width_range:
89
+ - 0
90
+ - 12
91
+ num_time_mask: 1
92
+ init_param: /root/.cache/modelscope/hub/iic/SenseVoiceSmall/model.pt
93
+ config: /root/.cache/modelscope/hub/iic/SenseVoiceSmall/config.yaml
94
+ is_training: true
95
+ trust_remote_code: true
96
+ train_data_set_list: /root/autodl-tmp/train/train.jsonl
97
+ valid_data_set_list: /root/autodl-tmp/val/val.jsonl
98
+ output_dir: ./outputs
99
+ model_path: /root/.cache/modelscope/hub/iic/SenseVoiceSmall
100
+ device: cpu
configuration.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "framework": "pytorch",
3
+ "task" : "auto-speech-recognition",
4
+ "model": {"type" : "funasr"},
5
+ "pipeline": {"type":"funasr-pipeline"},
6
+ "model_name_in_hub": {
7
+ "ms":"",
8
+ "hf":""},
9
+ "file_path_metas": {
10
+ "init_param":"model.pt",
11
+ "config":"config.yaml",
12
+ "tokenizer_conf": {"bpemodel": "chn_jpn_yue_eng_ko_spectok.bpe.model"},
13
+ "frontend_conf":{"cmvn_file": "am.mvn"}}
14
+ }
log.txt ADDED
File without changes
model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10fbed6aefd3504b48631b3060874b09a57f821c01111ab008a6fc32a1565368
3
+ size 2809076230
model.pt.best ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39da396056cc8340fe784419407a7266ae2d3529196b958d38460f80ce02cdf0
3
+ size 2809131854
tensorboard/events.out.tfevents.1729851842.autodl-container-f42f45a886-bccfcaff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61cd5645cb0a0457120b60d539673b4782db7d60b033f87cfa59bee0d82aafa3
3
+ size 37220360