Add required alphabet path to CI tests
This commit is contained in:
parent
920e92d68a
commit
5b4fa27467
@ -14,7 +14,7 @@ fi;
|
|||||||
# and when trying to run on multiple devices (like GPUs), this will break
|
# and when trying to run on multiple devices (like GPUs), this will break
|
||||||
export CUDA_VISIBLE_DEVICES=0
|
export CUDA_VISIBLE_DEVICES=0
|
||||||
|
|
||||||
python -u train.py --show_progressbar false --early_stop false \
|
python -u train.py --alphabet_config_path "data/alphabet.txt" --show_progressbar false --early_stop false \
|
||||||
--train_files ${ldc93s1_csv} --train_batch_size 1 \
|
--train_files ${ldc93s1_csv} --train_batch_size 1 \
|
||||||
--scorer "" \
|
--scorer "" \
|
||||||
--augment dropout \
|
--augment dropout \
|
||||||
|
@ -14,7 +14,7 @@ fi;
|
|||||||
# and when trying to run on multiple devices (like GPUs), this will break
|
# and when trying to run on multiple devices (like GPUs), this will break
|
||||||
export CUDA_VISIBLE_DEVICES=0
|
export CUDA_VISIBLE_DEVICES=0
|
||||||
|
|
||||||
python -u train.py --show_progressbar false --early_stop false \
|
python -u train.py --alphabet_config_path "data/alphabet.txt" --show_progressbar false --early_stop false \
|
||||||
--train_files ${ldc93s1_csv} --train_batch_size 1 \
|
--train_files ${ldc93s1_csv} --train_batch_size 1 \
|
||||||
--dev_files ${ldc93s1_csv} --dev_batch_size 1 \
|
--dev_files ${ldc93s1_csv} --dev_batch_size 1 \
|
||||||
--test_files ${ldc93s1_csv} --test_batch_size 1 \
|
--test_files ${ldc93s1_csv} --test_batch_size 1 \
|
||||||
|
@ -14,7 +14,7 @@ fi;
|
|||||||
# and when trying to run on multiple devices (like GPUs), this will break
|
# and when trying to run on multiple devices (like GPUs), this will break
|
||||||
export CUDA_VISIBLE_DEVICES=0
|
export CUDA_VISIBLE_DEVICES=0
|
||||||
|
|
||||||
python -u train.py --show_progressbar false --early_stop false \
|
python -u train.py --alphabet_config_path "data/alphabet.txt" --show_progressbar false --early_stop false \
|
||||||
--train_files ${ldc93s1_csv} --train_batch_size 1 \
|
--train_files ${ldc93s1_csv} --train_batch_size 1 \
|
||||||
--dev_files ${ldc93s1_csv} --dev_batch_size 1 \
|
--dev_files ${ldc93s1_csv} --dev_batch_size 1 \
|
||||||
--test_files ${ldc93s1_csv} --test_batch_size 1 \
|
--test_files ${ldc93s1_csv} --test_batch_size 1 \
|
||||||
|
@ -20,7 +20,7 @@ fi;
|
|||||||
# and when trying to run on multiple devices (like GPUs), this will break
|
# and when trying to run on multiple devices (like GPUs), this will break
|
||||||
export CUDA_VISIBLE_DEVICES=0
|
export CUDA_VISIBLE_DEVICES=0
|
||||||
|
|
||||||
python -u train.py --show_progressbar false --early_stop false \
|
python -u train.py --alphabet_config_path "data/alphabet.txt" --show_progressbar false --early_stop false \
|
||||||
--train_files ${ldc93s1_sdb} --train_batch_size 1 \
|
--train_files ${ldc93s1_sdb} --train_batch_size 1 \
|
||||||
--dev_files ${ldc93s1_sdb} --dev_batch_size 1 \
|
--dev_files ${ldc93s1_sdb} --dev_batch_size 1 \
|
||||||
--test_files ${ldc93s1_sdb} --test_batch_size 1 \
|
--test_files ${ldc93s1_sdb} --test_batch_size 1 \
|
||||||
|
@ -17,7 +17,7 @@ fi;
|
|||||||
# and when trying to run on multiple devices (like GPUs), this will break
|
# and when trying to run on multiple devices (like GPUs), this will break
|
||||||
export CUDA_VISIBLE_DEVICES=0
|
export CUDA_VISIBLE_DEVICES=0
|
||||||
|
|
||||||
python -u train.py --show_progressbar false --early_stop false \
|
python -u train.py --alphabet_config_path "data/alphabet.txt" --show_progressbar false --early_stop false \
|
||||||
--train_files ${ldc93s1_csv} --train_batch_size 1 \
|
--train_files ${ldc93s1_csv} --train_batch_size 1 \
|
||||||
--feature_cache '/tmp/ldc93s1_cache' \
|
--feature_cache '/tmp/ldc93s1_cache' \
|
||||||
--dev_files ${ldc93s1_csv} --dev_batch_size 1 \
|
--dev_files ${ldc93s1_csv} --dev_batch_size 1 \
|
||||||
|
@ -17,7 +17,7 @@ fi;
|
|||||||
# and when trying to run on multiple devices (like GPUs), this will break
|
# and when trying to run on multiple devices (like GPUs), this will break
|
||||||
export CUDA_VISIBLE_DEVICES=0
|
export CUDA_VISIBLE_DEVICES=0
|
||||||
|
|
||||||
python -u train.py --show_progressbar false --early_stop false \
|
python -u train.py --alphabet_config_path "data/alphabet.txt" --show_progressbar false --early_stop false \
|
||||||
--train_files ${ldc93s1_csv} --train_batch_size 1 \
|
--train_files ${ldc93s1_csv} --train_batch_size 1 \
|
||||||
--feature_cache '/tmp/ldc93s1_cache' \
|
--feature_cache '/tmp/ldc93s1_cache' \
|
||||||
--dev_files ${ldc93s1_csv} --dev_batch_size 1 \
|
--dev_files ${ldc93s1_csv} --dev_batch_size 1 \
|
||||||
|
@ -16,7 +16,7 @@ fi;
|
|||||||
# and when trying to run on multiple devices (like GPUs), this will break
|
# and when trying to run on multiple devices (like GPUs), this will break
|
||||||
export CUDA_VISIBLE_DEVICES=0
|
export CUDA_VISIBLE_DEVICES=0
|
||||||
|
|
||||||
python -u train.py --show_progressbar false \
|
python -u train.py --alphabet_config_path "data/alphabet.txt" --show_progressbar false \
|
||||||
--n_hidden 100 \
|
--n_hidden 100 \
|
||||||
--checkpoint_dir '/tmp/ckpt_bytes' \
|
--checkpoint_dir '/tmp/ckpt_bytes' \
|
||||||
--export_dir '/tmp/train_bytes_tflite' \
|
--export_dir '/tmp/train_bytes_tflite' \
|
||||||
|
@ -17,7 +17,7 @@ fi;
|
|||||||
# and when trying to run on multiple devices (like GPUs), this will break
|
# and when trying to run on multiple devices (like GPUs), this will break
|
||||||
export CUDA_VISIBLE_DEVICES=0
|
export CUDA_VISIBLE_DEVICES=0
|
||||||
|
|
||||||
python -u train.py --show_progressbar false --early_stop false \
|
python -u train.py --alphabet_config_path "data/alphabet.txt" --show_progressbar false --early_stop false \
|
||||||
--train_files ${ldc93s1_csv} --train_batch_size 1 \
|
--train_files ${ldc93s1_csv} --train_batch_size 1 \
|
||||||
--dev_files ${ldc93s1_csv} --dev_batch_size 1 \
|
--dev_files ${ldc93s1_csv} --dev_batch_size 1 \
|
||||||
--test_files ${ldc93s1_csv} --test_batch_size 1 \
|
--test_files ${ldc93s1_csv} --test_batch_size 1 \
|
||||||
|
@ -23,7 +23,7 @@ fi;
|
|||||||
# and when trying to run on multiple devices (like GPUs), this will break
|
# and when trying to run on multiple devices (like GPUs), this will break
|
||||||
export CUDA_VISIBLE_DEVICES=0
|
export CUDA_VISIBLE_DEVICES=0
|
||||||
|
|
||||||
python -u train.py --show_progressbar false --early_stop false \
|
python -u train.py --alphabet_config_path "data/alphabet.txt" --show_progressbar false --early_stop false \
|
||||||
--train_files ${ldc93s1_sdb} --train_batch_size 1 \
|
--train_files ${ldc93s1_sdb} --train_batch_size 1 \
|
||||||
--dev_files ${ldc93s1_sdb} --dev_batch_size 1 \
|
--dev_files ${ldc93s1_sdb} --dev_batch_size 1 \
|
||||||
--test_files ${ldc93s1_sdb} --test_batch_size 1 \
|
--test_files ${ldc93s1_sdb} --test_batch_size 1 \
|
||||||
|
@ -23,7 +23,7 @@ fi;
|
|||||||
# and when trying to run on multiple devices (like GPUs), this will break
|
# and when trying to run on multiple devices (like GPUs), this will break
|
||||||
export CUDA_VISIBLE_DEVICES=0
|
export CUDA_VISIBLE_DEVICES=0
|
||||||
|
|
||||||
python -u train.py --show_progressbar false --early_stop false \
|
python -u train.py --alphabet_config_path "data/alphabet.txt" --show_progressbar false --early_stop false \
|
||||||
--train_files ${ldc93s1_sdb} ${ldc93s1_csv} --train_batch_size 1 \
|
--train_files ${ldc93s1_sdb} ${ldc93s1_csv} --train_batch_size 1 \
|
||||||
--feature_cache '/tmp/ldc93s1_cache_sdb_csv' \
|
--feature_cache '/tmp/ldc93s1_cache_sdb_csv' \
|
||||||
--dev_files ${ldc93s1_sdb} ${ldc93s1_csv} --dev_batch_size 1 \
|
--dev_files ${ldc93s1_sdb} ${ldc93s1_csv} --dev_batch_size 1 \
|
||||||
|
@ -14,7 +14,7 @@ fi;
|
|||||||
# and when trying to run on multiple devices (like GPUs), this will break
|
# and when trying to run on multiple devices (like GPUs), this will break
|
||||||
export CUDA_VISIBLE_DEVICES=0
|
export CUDA_VISIBLE_DEVICES=0
|
||||||
|
|
||||||
python -u train.py --show_progressbar false --early_stop false \
|
python -u train.py --alphabet_config_path "data/alphabet.txt" --show_progressbar false --early_stop false \
|
||||||
--train_files ${ldc93s1_csv} --train_batch_size 1 \
|
--train_files ${ldc93s1_csv} --train_batch_size 1 \
|
||||||
--dev_files ${ldc93s1_csv} --dev_batch_size 1 \
|
--dev_files ${ldc93s1_csv} --dev_batch_size 1 \
|
||||||
--test_files ${ldc93s1_csv} --test_batch_size 1 \
|
--test_files ${ldc93s1_csv} --test_batch_size 1 \
|
||||||
@ -23,7 +23,7 @@ python -u train.py --show_progressbar false --early_stop false \
|
|||||||
--learning_rate 0.001 --dropout_rate 0.05 \
|
--learning_rate 0.001 --dropout_rate 0.05 \
|
||||||
--scorer_path 'data/smoke_test/pruned_lm.scorer'
|
--scorer_path 'data/smoke_test/pruned_lm.scorer'
|
||||||
|
|
||||||
python -u train.py \
|
python -u train.py --alphabet_config_path "data/alphabet.txt" \
|
||||||
--n_hidden 100 \
|
--n_hidden 100 \
|
||||||
--checkpoint_dir '/tmp/ckpt' \
|
--checkpoint_dir '/tmp/ckpt' \
|
||||||
--scorer_path 'data/smoke_test/pruned_lm.scorer' \
|
--scorer_path 'data/smoke_test/pruned_lm.scorer' \
|
||||||
|
@ -16,7 +16,7 @@ fi;
|
|||||||
# and when trying to run on multiple devices (like GPUs), this will break
|
# and when trying to run on multiple devices (like GPUs), this will break
|
||||||
export CUDA_VISIBLE_DEVICES=0
|
export CUDA_VISIBLE_DEVICES=0
|
||||||
|
|
||||||
python -u train.py --show_progressbar false \
|
python -u train.py --alphabet_config_path "data/alphabet.txt" --show_progressbar false \
|
||||||
--n_hidden 100 \
|
--n_hidden 100 \
|
||||||
--checkpoint_dir '/tmp/ckpt' \
|
--checkpoint_dir '/tmp/ckpt' \
|
||||||
--export_dir '/tmp/train_tflite' \
|
--export_dir '/tmp/train_tflite' \
|
||||||
@ -26,7 +26,7 @@ python -u train.py --show_progressbar false \
|
|||||||
|
|
||||||
mkdir /tmp/train_tflite/en-us
|
mkdir /tmp/train_tflite/en-us
|
||||||
|
|
||||||
python -u train.py --show_progressbar false \
|
python -u train.py --alphabet_config_path "data/alphabet.txt" --show_progressbar false \
|
||||||
--n_hidden 100 \
|
--n_hidden 100 \
|
||||||
--checkpoint_dir '/tmp/ckpt' \
|
--checkpoint_dir '/tmp/ckpt' \
|
||||||
--export_dir '/tmp/train_tflite/en-us' \
|
--export_dir '/tmp/train_tflite/en-us' \
|
||||||
|
@ -20,7 +20,7 @@ fi
|
|||||||
# and when trying to run on multiple devices (like GPUs), this will break
|
# and when trying to run on multiple devices (like GPUs), this will break
|
||||||
export CUDA_VISIBLE_DEVICES=0
|
export CUDA_VISIBLE_DEVICES=0
|
||||||
|
|
||||||
python -u train.py --show_progressbar false \
|
python -u train.py --alphabet_config_path "data/alphabet.txt" --show_progressbar false \
|
||||||
--train_files data/ldc93s1/ldc93s1.csv \
|
--train_files data/ldc93s1/ldc93s1.csv \
|
||||||
--test_files data/ldc93s1/ldc93s1.csv \
|
--test_files data/ldc93s1/ldc93s1.csv \
|
||||||
--train_batch_size 1 \
|
--train_batch_size 1 \
|
||||||
|
Loading…
Reference in New Issue
Block a user