update
Browse files
examples/fsmn_vad_by_webrtcvad/yaml/config-240-80-4.yaml
CHANGED
|
@@ -8,7 +8,7 @@ hop_size: 80
|
|
| 8 |
win_type: hann
|
| 9 |
|
| 10 |
# model
|
| 11 |
-
fsmn_input_size:
|
| 12 |
fsmn_input_affine_size: 140
|
| 13 |
fsmn_hidden_size: 250
|
| 14 |
fsmn_basic_block_layers: 4
|
|
|
|
| 8 |
win_type: hann
|
| 9 |
|
| 10 |
# model
|
| 11 |
+
fsmn_input_size: 129
|
| 12 |
fsmn_input_affine_size: 140
|
| 13 |
fsmn_hidden_size: 250
|
| 14 |
fsmn_basic_block_layers: 4
|
examples/fsmn_vad_by_webrtcvad/yaml/config-256-128-4.yaml
CHANGED
|
@@ -8,7 +8,7 @@ hop_size: 128
|
|
| 8 |
win_type: hann
|
| 9 |
|
| 10 |
# model
|
| 11 |
-
fsmn_input_size:
|
| 12 |
fsmn_input_affine_size: 140
|
| 13 |
fsmn_hidden_size: 250
|
| 14 |
fsmn_basic_block_layers: 4
|
|
|
|
| 8 |
win_type: hann
|
| 9 |
|
| 10 |
# model
|
| 11 |
+
fsmn_input_size: 129
|
| 12 |
fsmn_input_affine_size: 140
|
| 13 |
fsmn_hidden_size: 250
|
| 14 |
fsmn_basic_block_layers: 4
|
examples/fsmn_vad_by_webrtcvad/yaml/config-512-256-4.yaml
CHANGED
|
@@ -8,7 +8,7 @@ hop_size: 256
|
|
| 8 |
win_type: hann
|
| 9 |
|
| 10 |
# model
|
| 11 |
-
fsmn_input_size:
|
| 12 |
fsmn_input_affine_size: 140
|
| 13 |
fsmn_hidden_size: 250
|
| 14 |
fsmn_basic_block_layers: 4
|
|
|
|
| 8 |
win_type: hann
|
| 9 |
|
| 10 |
# model
|
| 11 |
+
fsmn_input_size: 257
|
| 12 |
fsmn_input_affine_size: 140
|
| 13 |
fsmn_hidden_size: 250
|
| 14 |
fsmn_basic_block_layers: 4
|
toolbox/torchaudio/models/vad/fsmn_vad/modeling_fsmn_vad.py
CHANGED
|
@@ -282,7 +282,14 @@ class FSMNVadModelExport(nn.Module):
|
|
| 282 |
|
| 283 |
|
| 284 |
def main1():
|
| 285 |
-
config = FSMNVadConfig(
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 286 |
model = FSMNVadPretrainedModel(config=config)
|
| 287 |
|
| 288 |
noisy = torch.randn(size=(1, 16000), dtype=torch.float32)
|
|
@@ -349,4 +356,4 @@ def main2():
|
|
| 349 |
|
| 350 |
|
| 351 |
if __name__ == "__main__":
|
| 352 |
-
|
|
|
|
| 282 |
|
| 283 |
|
| 284 |
def main1():
|
| 285 |
+
config = FSMNVadConfig(
|
| 286 |
+
nfft=256,
|
| 287 |
+
win_size=256,
|
| 288 |
+
hop_size=128,
|
| 289 |
+
fsmn_input_size=129,
|
| 290 |
+
fsmn_input_affine_size=140,
|
| 291 |
+
fsmn_hidden_size=250
|
| 292 |
+
)
|
| 293 |
model = FSMNVadPretrainedModel(config=config)
|
| 294 |
|
| 295 |
noisy = torch.randn(size=(1, 16000), dtype=torch.float32)
|
|
|
|
| 356 |
|
| 357 |
|
| 358 |
if __name__ == "__main__":
|
| 359 |
+
main1()
|