6640cf2341
* Redo remote I/O changes once more; this time without messing with taskcluster * Add bin changes * Fix merge-induced issue? * For the interleaved case with multiple collections, unpack audio on the fly To reproduce the previous failure rm data/smoke_test/ldc93s1.csv rm data/smoke_test/ldc93s1.sdb rm -rf /tmp/ldc93s1_cache_sdb_csv rm -rf /tmp/ckpt_sdb_csv rm -rf /tmp/train_sdb_csv ./bin/run-tc-ldc93s1_new_sdb_csv.sh 109 16000 python -u DeepSpeech.py --noshow_progressbar --noearly_stop --train_files ./data/smoke_test/ldc93s1.sdb,./data/smoke_test/ldc93s1.csv --train_batch_size 1 --feature_cache /tmp/ldc93s1_cache_sdb_csv --dev_files ./data/smoke_test/ldc93s1.sdb,./data/smoke_test/ldc93s1.csv --dev_batch_size 1 --test_files ./data/smoke_test/ldc93s1.sdb,./data/smoke_test/ldc93s1.csv --test_batch_size 1 --n_hidden 100 --epochs 109 --max_to_keep 1 --checkpoint_dir /tmp/ckpt_sdb_csv --learning_rate 0.001 --dropout_rate 0.05 --export_dir /tmp/train_sdb_csv --scorer_path data/smoke_test/pruned_lm.scorer --audio_sample_rate 16000 * Attempt to preserve length information with a wrapper around `map()`… this gets pretty python-y * Call the right `__next__()` * Properly implement the rest of the map wrappers here…… * Fix trailing whitespace situation and other linter complaints * Remove data accidentally checked in * Fix overlay augmentations * Wavs must be open in rb mode if we're passing in an external file pointer -- this confused me * Lint whitespace * Revert "Fix trailing whitespace situation and other linter complaints" This reverts commit c3c45397a2f98e9b00d00c18c4ced4fc52475032. * Fix linter issue but without such an aggressive diff * Move unpack_maybe into sample_collections * Use unpack_maybe in place of duplicate lambda * Fix confusing comment * Add clarifying comment for on-the-fly unpacking |
||
---|---|---|
.. | ||
compare_samples.py | ||
data_set_tool.py | ||
graphdef_binary_to_text.py | ||
import_aidatatang.py | ||
import_aishell.py | ||
import_ccpmf.py | ||
import_cv2.py | ||
import_cv.py | ||
import_fisher.py | ||
import_freestmandarin.py | ||
import_gram_vaani.py | ||
import_ldc93s1.py | ||
import_librivox.py | ||
import_lingua_libre.py | ||
import_m-ailabs.py | ||
import_magicdata.py | ||
import_primewords.py | ||
import_slr57.py | ||
import_swb.py | ||
import_swc.py | ||
import_ted.py | ||
import_timit.py | ||
import_ts.py | ||
import_tuda.py | ||
import_vctk.py | ||
import_voxforge.py | ||
ops_in_graph.py | ||
play.py | ||
README.rst | ||
run-ldc93s1.sh | ||
run-tc-graph_augmentations.sh | ||
run-tc-ldc93s1_checkpoint_bytes.sh | ||
run-tc-ldc93s1_checkpoint_sdb.sh | ||
run-tc-ldc93s1_checkpoint.sh | ||
run-tc-ldc93s1_new_bytes_tflite.sh | ||
run-tc-ldc93s1_new_bytes.sh | ||
run-tc-ldc93s1_new_metrics.sh | ||
run-tc-ldc93s1_new_sdb_csv.sh | ||
run-tc-ldc93s1_new_sdb.sh | ||
run-tc-ldc93s1_new.sh | ||
run-tc-ldc93s1_singleshotinference.sh | ||
run-tc-ldc93s1_tflite.sh | ||
run-tc-sample_augmentations.sh | ||
run-tc-transfer.sh |
Utility scripts =============== This folder contains scripts that can be used to do training on the various included importers from the command line. This is useful to be able to run training without a browser open, or unattended on a remote machine. They should be run from the base directory of the repository. Note that the default settings assume a very well-specified machine. In the situation that out-of-memory errors occur, you may find decreasing the values of ``--train_batch_size``\ , ``--dev_batch_size`` and ``--test_batch_size`` will allow you to continue, at the expense of speed.