Normalize sample rate of dev/test sets to avoid feature computation errors
This commit is contained in:
parent
c78af058a5
commit
8cdaa18533
@ -438,6 +438,7 @@ def train():
|
||||
dev_sets = [create_dataset([source],
|
||||
batch_size=FLAGS.dev_batch_size,
|
||||
train_phase=False,
|
||||
augmentations=[NormalizeSampleRate(FLAGS.audio_sample_rate)],
|
||||
exception_box=exception_box,
|
||||
process_ahead=len(Config.available_devices) * FLAGS.dev_batch_size * 2,
|
||||
reverse=FLAGS.reverse_dev,
|
||||
@ -450,6 +451,7 @@ def train():
|
||||
metrics_sets = [create_dataset([source],
|
||||
batch_size=FLAGS.dev_batch_size,
|
||||
train_phase=False,
|
||||
augmentations=[NormalizeSampleRate(FLAGS.audio_sample_rate)],
|
||||
exception_box=exception_box,
|
||||
process_ahead=len(Config.available_devices) * FLAGS.dev_batch_size * 2,
|
||||
reverse=FLAGS.reverse_dev,
|
||||
|
Loading…
Reference in New Issue
Block a user