trunk: updates to scripts to more consistenly use new-style queue.pl options (--num-threads, --mem, --gpu) and remove --parallel-opts wherever possible (work out the new-style option from --num-threads directly where relevant)

git-svn-id: https://svn.code.sf.net/p/kaldi/code/trunk@5104 5e6a8d80-dfce-4ca6-a32a-6e07a63d50c8
This commit is contained in:
Dan Povey 2015-05-31 22:51:42 +00:00
Родитель 06506962b4
Коммит 62690a7375
62 изменённых файлов: 204 добавлений и 257 удалений

Просмотреть файл

@ -67,10 +67,10 @@ if [ $stage -le 2 ]; then
$data_fmllr/$mic/train_tr90 $data_fmllr/$mic/train_cv10 data/lang $ali $ali $dir || exit 1;
# Decode (reuse HCLG graph)
steps/nnet/decode.sh --nj 6 --cmd "$decode_cmd" --config conf/decode_dnn.conf --acwt 0.1 \
--num-threads 3 --parallel-opts "-pe smp 4" \
--num-threads 3 \
$graph_dir $data_fmllr/$mic/dev $dir/decode_dev_${lm_suffix} || exit 1;
steps/nnet/decode.sh --nj 6 --cmd "$decode_cmd" --config conf/decode_dnn.conf --acwt 0.1 \
--num-threads 3 --parallel-opts "-pe smp 4" \
--num-threads 3 \
$graph_dir $data_fmllr/$mic/eval $dir/decode_eval_${lm_suffix} || exit 1;
fi
@ -96,11 +96,11 @@ if [ $stage -le 4 ]; then
# Decode (reuse HCLG graph)
for ITER in 1; do
steps/nnet/decode.sh --nj 6 --cmd "$decode_cmd" --config conf/decode_dnn.conf \
--num-threads 3 --parallel-opts "-pe smp 4" \
--num-threads 3 \
--nnet $dir/${ITER}.nnet --acwt $acwt \
$graph_dir $data_fmllr/$mic/dev $dir/decode_dev_${lm_suffix} || exit 1;
steps/nnet/decode.sh --nj 6 --cmd "$decode_cmd" --config conf/decode_dnn.conf \
--num-threads 3 --parallel-opts "-pe smp 4" \
--num-threads 3 \
--nnet $dir/${ITER}.nnet --acwt $acwt \
$graph_dir $data_fmllr/$mic/eval $dir/decode_eval_${lm_suffix} || exit 1;
done
@ -126,11 +126,11 @@ if [ $stage -le 6 ]; then
# Decode (reuse HCLG graph)
for ITER in 1 2 3 4; do
steps/nnet/decode.sh --nj 6 --cmd "$decode_cmd" --config conf/decode_dnn.conf \
--num-threads 3 --parallel-opts "-pe smp 4" \
--num-threads 3 \
--nnet $dir/${ITER}.nnet --acwt $acwt \
$graph_dir $data_fmllr/$mic/dev $dir/decode_dev_${lm_suffix}_$ITER || exit 1;
steps/nnet/decode.sh --nj 6 --cmd "$decode_cmd" --config conf/decode_dnn.conf \
--num-threads 3 --parallel-opts "-pe smp 4" \
--num-threads 3 \
--nnet $dir/${ITER}.nnet --acwt $acwt \
$graph_dir $data_fmllr/$mic/eval $dir/decode_eval_${lm_suffix}_$ITER || exit 1;
done

Просмотреть файл

@ -7,10 +7,10 @@
#a) JHU cluster options
export train_cmd="queue.pl -l arch=*64"
export decode_cmd="queue.pl -l arch=*64,mem_free=2G,ram_free=2G"
export mkgraph_cmd="queue.pl -l arch=*64,ram_free=4G,mem_free=4G"
export big_memory_cmd="queue.pl -l arch=*64,ram_free=8G,mem_free=8G"
export cuda_cmd="queue.pl -l gpu=1"
export decode_cmd="queue.pl -l arch=*64 --mem 2G"
export mkgraph_cmd="queue.pl -l arch=*64 --mem 4G"
export big_memory_cmd="queue.pl -l arch=*64 --mem 8G"
export cuda_cmd="queue.pl --gpu 1"
#b) BUT cluster options

Просмотреть файл

@ -33,7 +33,7 @@ lat_weights=
word_ins_penalty=0.0
min_lmwt=7
max_lmwt=17
parallel_opts="-pe smp 3"
parallel_opts="--num-threads 3"
skip_scoring=false
ctm_name=
#end configuration section.
@ -49,7 +49,7 @@ Options:
--cmd (run.pl|queue.pl...) # specify how to run the sub-processes.
--stage (0|1|2) # (createCTM | filterCTM | runSclite).
--parallel-opts <string> # extra options to command for combination stage,
# default '-pe smp 3'
# default '--num-threads 3'
--cer (0|1) # compute CER in addition to WER
";

Просмотреть файл

@ -13,7 +13,7 @@ set -u
if [ ! -f exp/combine_2/decode_dev2h/.done ]; then
for iter in 1 2 3 4; do
local/score_combine.sh --cmd "queue.pl -l mem_free=2.0G,ram_free=2.0G" \
local/score_combine.sh --cmd "queue.pl --mem 2G" \
data/dev2h data/lang exp/tri6_nnet/decode_dev2h exp/sgmm5_mmi_b0.1/decode_dev2h_fmllr_it$iter exp/combine_2/decode_dev2h_it$iter
touch exp/combine_2/decode_dev2h/.done
done
@ -25,7 +25,7 @@ if [ ! -f exp/combine_3/decode_dev2h/.done ]; then
if [ ! -f exp_BNF/sgmm7_mmi_b0.1/decode_dev2h_fmllr_it$iter/.done ]; then
echo "BNF decode in exp_BNF/sgmm7_mmi_b0.1/decode_dev2h_fmllr_it$iter is not done, skipping this step."
fi
local/score_combine.sh --cmd "queue.pl -l mem_free=2.0G,ram_free=2.0G" \
local/score_combine.sh --cmd "queue.pl --mem 2G" \
data/dev2h data/lang exp_BNF/sgmm7_mmi_b0.1/decode_dev2h_fmllr_it$iter:10 \
exp/sgmm5_mmi_b0.1/decode_dev2h_fmllr_it$iter exp/tri5_nnet/decode_dev2h exp/combine_3/decode_dev2h_it$iter
touch exp_BNF/sgmm7_mmi_b0.1/decode_dev2h_fmllr_it$iter/.done

Просмотреть файл

@ -236,7 +236,7 @@ utils/mkgraph.sh data/lang_test exp/sgmm5 exp/sgmm5/graph
(
steps/decode_sgmm2.sh --nj 13 --cmd "$decode_cmd" --num-threads 5 --parallel-opts " -pe smp 5" \
steps/decode_sgmm2.sh --nj 13 --cmd "$decode_cmd" --num-threads 5 \
--config conf/decode.config --scoring-opts "--min-lmwt 8 --max-lmwt 16" --transform-dir exp/tri5a/decode_dev \
exp/sgmm5/graph data/dev exp/sgmm5/decode_dev
)&
@ -247,7 +247,7 @@ steps/align_sgmm2.sh \
data/train data/lang exp/sgmm5 exp/sgmm5_ali
steps/make_denlats_sgmm2.sh \
--nj 32 --sub-split 32 --num-threads 4 --parallel-opts "-pe smp 4"\
--nj 32 --sub-split 32 --num-threads 4 \
--beam 10.0 --lattice-beam 6 --cmd "$decode_cmd" --transform-dir exp/tri5a_ali \
data/train data/lang exp/sgmm5_ali exp/sgmm5_denlats
@ -262,7 +262,7 @@ steps/decode_fmllr_extra.sh --nj 13 --cmd "$decode_cmd" --num-threads 4 --parall
--config conf/decode.config --scoring-opts "--min-lmwt 8 --max-lmwt 12"\
exp/tri5a/graph data/dev exp/tri5a/decode_dev
utils/mkgraph.sh data/lang_test exp/sgmm5 exp/sgmm5/graph
steps/decode_sgmm2.sh --nj 13 --cmd "$decode_cmd" --num-threads 5 --parallel-opts " -pe smp 5" \
steps/decode_sgmm2.sh --nj 13 --cmd "$decode_cmd" --num-threads 5 \
--config conf/decode.config --scoring-opts "--min-lmwt 8 --max-lmwt 16" --transform-dir exp/tri5a/decode_dev \
exp/sgmm5/graph data/dev exp/sgmm5/decode_dev
for iter in 1 2 3 4; do
@ -276,9 +276,9 @@ done
dnn_cpu_parallel_opts=(--minibatch-size 128 --max-change 10 --num-jobs-nnet 8 --num-threads 16 \
--parallel-opts "-pe smp 16" --cmd "queue.pl -l arch=*64 -l mem_free=2G,ram_free=2G")
--parallel-opts "-pe smp 16" --cmd "queue.pl -l arch=*64 --mem 2G")
dnn_gpu_parallel_opts=(--minibatch-size 512 --max-change 40 --num-jobs-nnet 4 --num-threads 1 \
--parallel-opts "-l gpu=1" --cmd "queue.pl -l arch=*64 -l mem_free=2G,ram_free=2G")
--parallel-opts "-l gpu=1" --cmd "queue.pl -l arch=*64 --mem 2G")
steps/nnet2/train_pnorm_ensemble.sh \
--mix-up 5000 --initial-learning-rate 0.008 --final-learning-rate 0.0008\

Просмотреть файл

@ -7,9 +7,9 @@
#a) JHU cluster options
export train_cmd="queue.pl -l arch=*64"
export decode_cmd="queue.pl -l arch=*64,mem_free=2G,ram_free=2G"
export mkgraph_cmd="queue.pl -l arch=*64,ram_free=4G,mem_free=4G"
export big_memory_cmd="queue.pl -l arch=*64,ram_free=8G,mem_free=8G"
export decode_cmd="queue.pl -l arch=*64 --mem 2G"
export mkgraph_cmd="queue.pl -l arch=*64 --mem 4G"
export big_memory_cmd="queue.pl -l arch=*64 --mem 8G"
export cuda_cmd="queue.pl -l gpu=1"

Просмотреть файл

@ -7,9 +7,9 @@
#a) JHU cluster options
#export train_cmd="queue.pl -l arch=*64"
#export decode_cmd="queue.pl -l arch=*64,mem_free=2G,ram_free=2G"
#export mkgraph_cmd="queue.pl -l arch=*64,ram_free=4G,mem_free=4G"
#export big_memory_cmd="queue.pl -l arch=*64,ram_free=8G,mem_free=8G"
#export decode_cmd="queue.pl -l arch=*64 --mem 2G"
#export mkgraph_cmd="queue.pl -l arch=*64 --mem 2G"
#export big_memory_cmd="queue.pl -l arch=*64 --mem 2G"
#export cuda_cmd="queue.pl -l gpu=1"

Просмотреть файл

@ -7,9 +7,9 @@
#a) JHU cluster options
export train_cmd="queue.pl -l arch=*64*"
export decode_cmd="queue.pl -l arch=*64* -l ram_free=4G,mem_free=4G"
export decode_cmd="queue.pl -l arch=*64* --mem 4G"
#export cuda_cmd="..."
export mkgraph_cmd="queue.pl -l arch=*64* -l ram_free=4G,mem_free=4G"
export mkgraph_cmd="queue.pl -l arch=*64* --mem 4G"
#b) BUT cluster options
#export train_cmd="queue.pl -q all.q@@blade -l ram_free=1200M,mem_free=1200M"

Просмотреть файл

@ -12,8 +12,8 @@
# JHU cluster:
export train_cmd="queue.pl -l arch=*64*"
export decode_cmd="queue.pl -l arch=*64* -l ram_free=4G,mem_free=4G"
export cuda_cmd="queue.pl -l arch=*64*,gpu=1 -q g.q"
export decode_cmd="queue.pl -l arch=*64* --mem 4G"
export cuda_cmd="queue.pl -l arch=*64* --gpu 1"
host=$(hostname -f)
if [ ${host#*.} == "fit.vutbr.cz" ]; then

Просмотреть файл

@ -70,10 +70,10 @@ if [ $stage -le 3 ]; then
utils/mkgraph.sh data/lang_nosp_test exp/tri1 exp/tri1/graph_nosp || exit 1
steps/decode.sh --nj $decode_nj --cmd "$decode_cmd" \
--num-threads 4 --parallel-opts "--num-threads 4" \
--num-threads 4 \
exp/tri1/graph_nosp data/dev exp/tri1/decode_nosp_dev || exit 1
steps/decode.sh --nj $decode_nj --cmd "$decode_cmd" \
--num-threads 4 --parallel-opts "--num-threads 4" \
--num-threads 4 \
exp/tri1/graph_nosp data/test exp/tri1/decode_nosp_test || exit 1
fi
@ -87,10 +87,10 @@ if [ $stage -le 4 ]; then
utils/mkgraph.sh data/lang_nosp_test exp/tri2 exp/tri2/graph_nosp || exit 1
steps/decode.sh --nj $decode_nj --cmd "$decode_cmd" \
--num-threads 4 --parallel-opts "--num-threads 4" \
--num-threads 4 \
exp/tri2/graph_nosp data/dev exp/tri2/decode_nosp_dev || exit 1
steps/decode.sh --nj $decode_nj --cmd "$decode_cmd" \
--num-threads 4 --parallel-opts "--num-threads 4" \
--num-threads 4 \
exp/tri2/graph_nosp data/test exp/tri2/decode_nosp_test || exit 1
fi
@ -110,10 +110,10 @@ if [ $stage -le 5 ]; then
utils/mkgraph.sh data/lang_test exp/tri2 exp/tri2/graph || exit 1
steps/decode.sh --nj $decode_nj --cmd "$decode_cmd" \
--num-threads 4 --parallel-opts "--num-threads 4" \
--num-threads 4 \
exp/tri2/graph data/dev exp/tri2/decode_dev || exit 1
steps/decode.sh --nj $decode_nj --cmd "$decode_cmd" \
--num-threads 4 --parallel-opts "--num-threads 4" \
--num-threads 4 \
exp/tri2/graph data/test exp/tri2/decode_test || exit 1
fi
@ -127,10 +127,10 @@ if [ $stage -le 6 ]; then
utils/mkgraph.sh data/lang_test exp/tri3 exp/tri3/graph || exit 1
steps/decode_fmllr.sh --nj $decode_nj --cmd "$decode_cmd" \
--num-threads 4 --parallel-opts "--num-threads 4" \
--num-threads 4 \
exp/tri3/graph data/dev exp/tri3/decode_dev || exit 1
steps/decode_fmllr.sh --nj $decode_nj --cmd "$decode_cmd" \
--num-threads 4 --parallel-opts "--num-threads 4" \
--num-threads 4 \
exp/tri3/graph data/test exp/tri3/decode_test || exit 1
fi
@ -147,10 +147,10 @@ if [ $stage -le 7 ]; then
for iter in 4; do
steps/decode.sh --transform-dir exp/tri3/decode_dev --nj $decode_nj --cmd "$decode_cmd" --iter $iter \
--num-threads 4 --parallel-opts "--num-threads 4" \
--num-threads 4 \
exp/tri3/graph data/dev exp/tri3_mmi_b0.1/decode_dev_it$iter || exit 1
steps/decode.sh --transform-dir exp/tri3/decode_test --nj $decode_nj --cmd "$decode_cmd" --iter $iter \
--num-threads 4 --parallel-opts "--num-threads 4" \
--num-threads 4 \
exp/tri3/graph data/test exp/tri3_mmi_b0.1/decode_test_it$iter || exit 1
done
fi

Просмотреть файл

@ -14,8 +14,8 @@
if [[ $(hostname -f) == *.clsp.jhu.edu ]]; then
export train_cmd="queue.pl -l arch=*64*"
export decode_cmd="queue.pl -l arch=*64*,ram_free=3G,mem_free=3G"
export mkgraph_cmd="queue.pl -l arch=*64*,ram_free=4G,mem_free=4G"
export decode_cmd="queue.pl -l arch=*64* --mem 3G"
export mkgraph_cmd="queue.pl -l arch=*64* --mem 4G"
export cuda_cmd="queue.pl -l gpu=1"
elif [[ $(hostname -f) == *.fit.vutbr.cz ]]; then
#b) BUT cluster options

Просмотреть файл

@ -191,7 +191,7 @@ echo " DNN Hybrid Training & Decoding
echo ============================================================================
# DNN hybrid system training parameters
dnn_mem_reqs="mem_free=1.0G,ram_free=1.0G"
dnn_mem_reqs="--mem 1G"
dnn_extra_opts="--num_epochs 20 --num-epochs-extra 10 --add-layers-period 1 --shrink-interval 3"
steps/nnet2/train_tanh.sh --mix-up 5000 --initial-learning-rate 0.015 \
@ -200,7 +200,7 @@ steps/nnet2/train_tanh.sh --mix-up 5000 --initial-learning-rate 0.015 \
data/train data/lang exp/tri3_ali exp/tri4_nnet
[ ! -d exp/tri4_nnet/decode_dev ] && mkdir -p exp/tri4_nnet/decode_dev
decode_extra_opts=(--num-threads 6 --parallel-opts "-pe smp 6 -l mem_free=4G,ram_free=4.0G")
decode_extra_opts=(--num-threads 6)
steps/nnet2/decode.sh --cmd "$decode_cmd" --nj "$decode_nj" "${decode_extra_opts[@]}" \
--transform-dir exp/tri3/decode_dev exp/tri3/graph data/dev \
exp/tri4_nnet/decode_dev | tee exp/tri4_nnet/decode_dev/decode.log

Просмотреть файл

@ -7,9 +7,9 @@
#a) JHU cluster options
export train_cmd="queue.pl -l arch=*64"
export decode_cmd="queue.pl -l arch=*64,mem_free=2G,ram_free=2G"
export mkgraph_cmd="queue.pl -l arch=*64,ram_free=4G,mem_free=4G"
export big_memory_cmd="queue.pl -l arch=*64,ram_free=8G,mem_free=8G"
export decode_cmd="queue.pl -l arch=*64 --mem 2G"
export mkgraph_cmd="queue.pl -l arch=*64 --mem 4G"
export big_memory_cmd="queue.pl -l arch=*64 --mem 8G"
export cuda_cmd="queue.pl -l gpu=1"

Просмотреть файл

@ -31,7 +31,7 @@ else
# almost the same, but this may be a little bit slow.
num_threads=16
minibatch_size=128
parallel_opts="-pe smp $num_threads"
parallel_opts="--num-threads $num_threads"
dir=$nnet_dir/nnet_a
fi

Просмотреть файл

@ -16,7 +16,7 @@ bnf_num_gauss_ubm=600
bnf_num_gauss_sgmm=7000
align_dir=exp/tri4b_ali_si284
bnf_decode_acwt=0.0357
sgmm_group_extra_opts=(--group 3 --parallel-opts "-pe smp 3 -l mem_free=7G,ram_free=7G" --cmd "queue.pl -l arch=*64 -l mem_free=2.0G,ram_free=2.0G")
sgmm_group_extra_opts=(--group 3 --cmd "queue.pl -l arch=*64 --mem 7G")
if [ ! -d exp_bnf ]; then
echo "$0: before running this script, please run local/run_bnf.sh"

Просмотреть файл

@ -51,7 +51,7 @@ local/wsj_format_data.sh --lang-suffix "_nosp" || exit 1;
# Note: I am commenting out the RNNLM-building commands below. They take up a lot
# of CPU time and are not really part of the "main recipe."
# Be careful: appending things like "-l mem_free=10G" to $decode_cmd
# Be careful: appending things like "--mem 10G" to $decode_cmd
# won't always work, it depends what $decode_cmd is.
(
local/wsj_extend_dict.sh --dict-suffix "_nosp" $wsj1/13-32.1 && \
@ -61,18 +61,18 @@ local/wsj_format_data.sh --lang-suffix "_nosp" || exit 1;
local/wsj_format_local_lms.sh --lang-suffix "_nosp" # &&
#
# ( local/wsj_train_rnnlms.sh --dict-suffix "_nosp" \
# --cmd "$decode_cmd -l mem_free=10G" data/local/rnnlm.h30.voc10k &
# --cmd "$decode_cmd --mem 10G" data/local/rnnlm.h30.voc10k &
# sleep 20; # wait till tools compiled.
# local/wsj_train_rnnlms.sh --dict-suffix "_nosp" \
# --cmd "$decode_cmd -l mem_free=12G" \
# --cmd "$decode_cmd --mem 12G" \
# --hidden 100 --nwords 20000 --class 350 \
# --direct 1500 data/local/rnnlm.h100.voc20k &
# local/wsj_train_rnnlms.sh --dict-suffix "_nosp" \
# --cmd "$decode_cmd -l mem_free=14G" \
# --cmd "$decode_cmd --mem 14G" \
# --hidden 200 --nwords 30000 --class 350 \
# --direct 1500 data/local/rnnlm.h200.voc30k &
# local/wsj_train_rnnlms.sh --dict-suffix "_nosp" \
# --cmd "$decode_cmd -l mem_free=16G" \
# --cmd "$decode_cmd --mem 16G" \
# --hidden 300 --nwords 40000 --class 400 \
# --direct 2000 data/local/rnnlm.h300.voc40k &
# )
@ -81,19 +81,19 @@ local/wsj_format_data.sh --lang-suffix "_nosp" || exit 1;
num_threads_rnnlm=8
local/wsj_train_rnnlms.sh --dict-suffix "_nosp" \
--rnnlm_ver rnnlm-hs-0.1b --threads $num_threads_rnnlm \
--cmd "$decode_cmd -l mem_free=1G --num-threads $num_threads_rnnlm" --bptt 4 --bptt-block 10 \
--cmd "$decode_cmd --mem 1G --num-threads $num_threads_rnnlm" --bptt 4 --bptt-block 10 \
--hidden 30 --nwords 10000 --direct 1000 data/local/rnnlm-hs.h30.voc10k
local/wsj_train_rnnlms.sh --dict-suffix "_nosp" \
--rnnlm_ver rnnlm-hs-0.1b --threads $num_threads_rnnlm \
--cmd "$decode_cmd -l mem_free=1G --num-threads $num_threads_rnnlm" --bptt 4 --bptt-block 10 \
--cmd "$decode_cmd --mem 1G --num-threads $num_threads_rnnlm" --bptt 4 --bptt-block 10 \
--hidden 100 --nwords 20000 --direct 1500 data/local/rnnlm-hs.h100.voc20k
local/wsj_train_rnnlms.sh --dict-suffix "_nosp" \
--rnnlm_ver rnnlm-hs-0.1b --threads $num_threads_rnnlm \
--cmd "$decode_cmd -l mem_free=1G --num-threads $num_threads_rnnlm" --bptt 4 --bptt-block 10 \
--cmd "$decode_cmd --mem 1G --num-threads $num_threads_rnnlm" --bptt 4 --bptt-block 10 \
--hidden 300 --nwords 30000 --direct 1500 data/local/rnnlm-hs.h300.voc30k
local/wsj_train_rnnlms.sh --dict-suffix "_nosp" \
--rnnlm_ver rnnlm-hs-0.1b --threads $num_threads_rnnlm \
--cmd "$decode_cmd -l mem_free=1G --num-threads $num_threads_rnnlm" --bptt 4 --bptt-block 10 \
--cmd "$decode_cmd --mem 1G --num-threads $num_threads_rnnlm" --bptt 4 --bptt-block 10 \
--hidden 400 --nwords 40000 --direct 2000 data/local/rnnlm-hs.h400.voc40k
)
) &

Просмотреть файл

@ -16,7 +16,7 @@ beam=13.0
lattice_beam=6.0
acwt=0.083333 # note: only really affects pruning (scoring is on lattices).
num_threads=1 # if >1, will use gmm-latgen-faster-parallel
parallel_opts= # If you supply num-threads, you should supply this too.
parallel_opts= # ignored now.
scoring_opts=
# note: there are no more min-lmwt and max-lmwt options, instead use
# e.g. --scoring-opts "--min-lmwt 1 --max-lmwt 20"
@ -48,7 +48,6 @@ if [ $# != 3 ]; then
echo " --acwt <float> # acoustic scale used for lattice generation "
echo " --scoring-opts <string> # options to local/score.sh"
echo " --num-threads <n> # number of threads to use, default 1."
echo " --parallel-opts <opts> # e.g. '-pe smp 4' if you supply --num-threads 4"
exit 1;
fi
@ -115,7 +114,7 @@ if [ $stage -le 0 ]; then
[ "`cat $graphdir/num_pdfs`" -eq `am-info --print-args=false $model | grep pdfs | awk '{print $NF}'` ] || \
{ echo "Mismatch in number of pdfs with $model"; exit 1; }
fi
$cmd $parallel_opts JOB=1:$nj $dir/log/decode.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/decode.JOB.log \
gmm-latgen-faster$thread_string --max-active=$max_active --beam=$beam --lattice-beam=$lattice_beam \
--acoustic-scale=$acwt --allow-partial=true --word-symbol-table=$graphdir/words.txt \
$model "$HCLG" "$feats" "ark:|gzip -c > $dir/lat.JOB.gz" || exit 1;

Просмотреть файл

@ -16,7 +16,7 @@ beam=13.0
lattice_beam=6.0
acwt=0.083333 # note: only really affects pruning (scoring is on lattices).
num_threads=1 # if >1, will use gmm-latgen-faster-parallel
parallel_opts= # If you supply num-threads, you should supply this too.
parallel_opts= # ignored now.
scoring_opts=
# note: there are no more min-lmwt and max-lmwt options, instead use
# e.g. --scoring-opts "--min-lmwt 1 --max-lmwt 20"
@ -48,7 +48,7 @@ if [ $# != 3 ]; then
echo " --acwt <float> # acoustic scale used for lattice generation "
echo " --scoring-opts <string> # options to local/score.sh"
echo " --num-threads <n> # number of threads to use, default 1."
echo " --parallel-opts <opts> # e.g. '-pe smp 4' if you supply --num-threads 4"
echo " --parallel-opts <opts> # ignored now, present for historical reasons."
exit 1;
fi
@ -110,7 +110,7 @@ if [ $stage -le 0 ]; then
[ "`cat $graphdir/num_pdfs`" -eq `am-info --print-args=false $model | grep pdfs | awk '{print $NF}'` ] || \
{ echo "Mismatch in number of pdfs with $model"; exit 1; }
fi
$cmd $parallel_opts JOB=1:$nj $dir/log/decode.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/decode.JOB.log \
gmm-latgen-faster$thread_string --max-active=$max_active --beam=$beam --lattice-beam=$lattice_beam \
--acoustic-scale=$acwt --allow-partial=true --word-symbol-table=$graphdir/words.txt \
$model $graphdir/HCLG.fst "$feats" "ark:|gzip -c > $dir/lat.JOB.gz" || exit 1;

Просмотреть файл

@ -49,7 +49,7 @@ silence_weight=0.01
cmd=run.pl
si_dir=
num_threads=1 # if >1, will use gmm-latgen-faster-parallel
parallel_opts= # If you supply num-threads, you should supply this too.
parallel_opts= # ignored, present for historical reasons.
skip_scoring=false
scoring_opts=
# End configuration section
@ -75,7 +75,7 @@ if [ $# != 3 ]; then
echo " --acwt <acoustic-weight> # default 0.08333 ... used to get posteriors"
echo " --scoring-opts <string> # options to local/score.sh"
echo " --num-threads <n> # number of threads to use, default 1."
echo " --parallel-opts <opts> # e.g. '-pe smp 4' if you supply --num-threads 4"
echo " --parallel-opts <opts> # ignored, present for historical reasons."
exit 1;
fi
@ -121,7 +121,7 @@ if [ -z "$si_dir" ]; then # we need to do the speaker-independent decoding pass.
{ echo "Mismatch in number of pdfs with $alignment_model"; exit 1; }
fi
steps/decode.sh --parallel-opts "$parallel_opts" --scoring-opts "$scoring_opts" \
steps/decode.sh --scoring-opts "$scoring_opts" \
--num-threads $num_threads --skip-scoring $skip_scoring \
--acwt $acwt --nj $nj --cmd "$cmd" --beam $first_beam \
--model $alignment_model --max-active \
@ -178,7 +178,7 @@ if [ $stage -le 2 ]; then
[ "`cat $graphdir/num_pdfs`" -eq `am-info --print-args=false $adapt_model | grep pdfs | awk '{print $NF}'` ] || \
{ echo "Mismatch in number of pdfs with $adapt_model"; exit 1; }
fi
$cmd JOB=1:$nj $dir/log/decode.JOB.log \
$cmd JOB=1:$nj --num-threads $num_threads $dir/log/decode.JOB.log \
gmm-latgen-faster$thread_string --max-active=$max_active --beam=$beam --lattice-beam=$lattice_beam \
--acoustic-scale=$acwt \
--determinize-lattice=false --allow-partial=true --word-symbol-table=$graphdir/words.txt \

Просмотреть файл

@ -1,6 +1,6 @@
#!/bin/bash
# Copyright 2012 Johns Hopkins University (Author: Daniel Povey)
# Copyright 2012-2015 Johns Hopkins University (Author: Daniel Povey)
# Decoding script that does fMLLR. This can be on top of delta+delta-delta, or
# LDA+MLLT features.
@ -42,7 +42,7 @@ cmd=run.pl
si_dir=
fmllr_update_type=full
num_threads=1 # if >1, will use gmm-latgen-faster-parallel
parallel_opts= # If you supply num-threads, you should supply this too.
parallel_opts= # ignored now.
skip_scoring=false
scoring_opts=
max_fmllr_jobs=25 # I've seen the fMLLR jobs overload NFS badly if the decoding
@ -69,7 +69,6 @@ if [ $# != 3 ]; then
echo " # Caution-- must be with same tree"
echo " --acwt <acoustic-weight> # default 0.08333 ... used to get posteriors"
echo " --num-threads <n> # number of threads to use, default 1."
echo " --parallel-opts <opts> # e.g. '-pe smp 4' if you supply --num-threads 4"
echo " --scoring-opts <opts> # options to local/score.sh"
exit 1;
fi
@ -117,11 +116,11 @@ if [ -z "$si_dir" ]; then # we need to do the speaker-independent decoding pass.
[ "`cat $graphdir/num_pdfs`" -eq `am-info --print-args=false $alignment_model | grep pdfs | awk '{print $NF}'` ] || \
{ echo "Mismatch in number of pdfs with $alignment_model"; exit 1; }
fi
steps/decode.sh --parallel-opts "$parallel_opts" --scoring-opts "$scoring_opts" \
--num-threads $num_threads --skip-scoring $skip_scoring \
--acwt $acwt --nj $nj --cmd "$cmd" --beam $first_beam \
--model $alignment_model --max-active \
$first_max_active $graphdir $data $si_dir || exit 1;
steps/decode.sh --scoring-opts "$scoring_opts" \
--num-threads $num_threads --skip-scoring $skip_scoring \
--acwt $acwt --nj $nj --cmd "$cmd" --beam $first_beam \
--model $alignment_model --max-active \
$first_max_active $graphdir $data $si_dir || exit 1;
fi
fi
##
@ -171,7 +170,7 @@ if [ $stage -le 2 ]; then
[ "`cat $graphdir/num_pdfs`" -eq `am-info --print-args=false $adapt_model | grep pdfs | awk '{print $NF}'` ] || \
{ echo "Mismatch in number of pdfs with $adapt_model"; exit 1; }
fi
$cmd $parallel_opts JOB=1:$nj $dir/log/decode.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/decode.JOB.log \
gmm-latgen-faster$thread_string --max-active=$max_active --beam=$beam --lattice-beam=$lattice_beam \
--acoustic-scale=$acwt --determinize-lattice=false \
--allow-partial=true --word-symbol-table=$graphdir/words.txt \
@ -208,7 +207,7 @@ feats="$sifeats transform-feats --utt2spk=ark:$sdata/JOB/utt2spk ark:$dir/trans.
if [ $stage -le 4 ]; then
echo "$0: doing a final pass of acoustic rescoring."
$cmd $parallel_opts JOB=1:$nj $dir/log/acoustic_rescore.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/acoustic_rescore.JOB.log \
gmm-rescore-lattice $final_model "ark:gunzip -c $dir/lat.tmp.JOB.gz|" "$feats" ark:- \| \
lattice-determinize-pruned$thread_string --acoustic-scale=$acwt --beam=$lattice_beam ark:- \
"ark:|gzip -c > $dir/lat.JOB.gz" '&&' rm $dir/lat.tmp.JOB.gz || exit 1;

Просмотреть файл

@ -56,7 +56,7 @@ si_dir=
fmllr_update_type=full
skip_scoring=false
num_threads=1 # if >1, will use gmm-latgen-faster-parallel
parallel_opts= # If you supply num-threads, you should supply this too.
parallel_opts= # ignored now.
scoring_opts=
# End configuration section
@ -81,7 +81,6 @@ if [ $# != 3 ]; then
echo " # Caution-- must be with same tree"
echo " --acwt <acoustic-weight> # default 0.08333 ... used to get posteriors"
echo " --num-threads <n> # number of threads to use, default 1."
echo " --parallel-opts <opts> # e.g. '-pe smp 4' if you supply --num-threads 4"
echo " --scoring-opts <opts> # options to local/score.sh"
exit 1;
fi
@ -129,7 +128,7 @@ if [ -z "$si_dir" ]; then # we need to do the speaker-independent decoding pass.
{ echo "Mismatch in number of pdfs with $alignment_model" exit 1; }
fi
steps/decode.sh --acwt $acwt --nj $nj --cmd "$cmd" --beam $first_beam --model $alignment_model\
--max-active $first_max_active --parallel-opts "${parallel_opts}" --num-threads $num_threads\
--max-active $first_max_active --num-threads $num_threads\
--skip-scoring true $graphdir $data $si_dir || exit 1;
fi
fi
@ -178,7 +177,7 @@ if [ $stage -le 2 ]; then
[ "`cat $graphdir/num_pdfs`" -eq `am-info --print-args=false $adapt_model | grep pdfs | awk '{print $NF}'` ] || \
{ echo "Mismatch in number of pdfs with $adapt_model" exit 1; }
fi
$cmd $parallel_opts JOB=1:$nj $dir/log/decode1.JOB.log\
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/decode1.JOB.log\
gmm-latgen-faster$thread_string --max-active=$first_max_active --max-mem=$max_mem --beam=$first_beam --lattice-beam=$first_lattice_beam \
--acoustic-scale=$acwt --allow-partial=true --word-symbol-table=$graphdir/words.txt \
$adapt_model $graphdir/HCLG.fst "$pass1feats" "ark:|gzip -c > $dir/lat1.JOB.gz" \
@ -214,7 +213,7 @@ pass2feats="$sifeats transform-feats --utt2spk=ark:$sdata/JOB/utt2spk ark:$dir/t
## after another stage of adaptation.)
if [ $stage -le 4 ]; then
echo "$0: doing final lattice generation phase"
$cmd $parallel_opts JOB=1:$nj $dir/log/decode2.JOB.log\
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/decode2.JOB.log\
gmm-latgen-faster$thread_string --max-active=$max_active --max-mem=$max_mem --beam=$beam --lattice-beam=$lattice_beam \
--acoustic-scale=$acwt --allow-partial=true --word-symbol-table=$graphdir/words.txt \
$adapt_model $graphdir/HCLG.fst "$pass2feats" "ark:|gzip -c > $dir/lat2.JOB.gz" \

Просмотреть файл

@ -17,7 +17,7 @@ acwt=0.083333 # note: only really affects pruning (scoring is on lattices).
ngselect=2; # Just use the 2 top Gaussians for fMMI/fMPE. Should match train.
transform_dir=
num_threads=1 # if >1, will use gmm-latgen-faster-parallel
parallel_opts= # If you supply num-threads, you should supply this too.
parallel_opts= # ignored now.
scoring_opts=
# End configuration section.
@ -46,7 +46,6 @@ if [ $# != 3 ]; then
echo " --scoring-opts <string> # options to local/score.sh"
echo " # speaker-adapted decoding"
echo " --num-threads <n> # number of threads to use, default 1."
echo " --parallel-opts <opts> # e.g. '-pe smp 4' if you supply --num-threads 4"
exit 1;
fi
@ -98,7 +97,7 @@ if [ $stage -le 1 ]; then
fi
if [ $stage -le 2 ]; then
$cmd $parallel_opts JOB=1:$nj $dir/log/decode.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/decode.JOB.log \
gmm-latgen-faster$thread_string --max-active=$maxactive --beam=$beam --lattice-beam=$lattice_beam \
--acoustic-scale=$acwt --allow-partial=true --word-symbol-table=$graphdir/words.txt \
$model $graphdir/HCLG.fst "$fmpefeats" "ark:|gzip -c > $dir/lat.JOB.gz" || exit 1;

Просмотреть файл

@ -18,7 +18,7 @@ logdet_scale=0.0
cmd=run.pl
skip_scoring=false
num_threads=1 # if >1, will use gmm-latgen-faster-parallel
parallel_opts= # If you supply num-threads, you should supply this too.
parallel_opts= # ignored now.
scoring_opts=
cleanup=true
# End configuration section
@ -89,7 +89,7 @@ if [ $stage -le 0 ]; then
[ "`cat $graphdir/num_pdfs`" -eq `am-info --print-args=false $srcdir/final.alimdl | grep pdfs | awk '{print $NF}'` ] || \
{ echo "Mismatch in number of pdfs with $srcdir/final.alimdl"; exit 1; }
fi
$cmd $parallel_opts JOB=1:$nj $dir/log/decode.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/decode.JOB.log \
gmm-latgen-faster$thread_string --max-active=$max_active --beam=$beam --lattice-beam=$lattice_beam \
--acoustic-scale=$acwt --allow-partial=true --word-symbol-table=$graphdir/words.txt \
$srcdir/final.alimdl $graphdir/HCLG.fst "$sifeats" "ark:|gzip -c > $dir/lat_pass1.JOB.gz" \
@ -121,7 +121,7 @@ feats1="$sifeats transform-feats --utt2spk=ark:$sdata/JOB/utt2spk ark:$dir/trans
if [ $stage -le 3 ]; then
echo "$0: rescoring the lattices with first-pass LVTLN transforms"
$cmd $parallel_opts JOB=1:$nj $dir/log/rescore.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/rescore.JOB.log \
gmm-rescore-lattice $srcdir/final.mdl "ark:gunzip -c $dir/lat_pass1.JOB.gz|" "$feats1" \
"ark:|gzip -c > $dir/lat_pass2.JOB.gz" || exit 1;
fi
@ -144,7 +144,7 @@ if [ $stage -le 5 ]; then
# This second rescoring is only really necessary for scoring purposes,
# it does not affect the transforms.
echo "$0: rescoring the lattices with second-pass LVTLN transforms"
$cmd $parallel_opts JOB=1:$nj $dir/log/rescore.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/rescore.JOB.log \
gmm-rescore-lattice $srcdir/final.mdl "ark:gunzip -c $dir/lat_pass2.JOB.gz|" "$feats" \
"ark:|gzip -c > $dir/lat.JOB.gz" || exit 1;
fi

Просмотреть файл

@ -23,10 +23,6 @@ max_active=7000
beam=13.0
lattice_beam=6.0
acwt=0.083333 # note: only really affects pruning (scoring is on lattices).
num_threads=1 # if >1, will use gmm-latgen-faster-parallel
parallel_opts= # If you supply num-threads, you should supply this too.
# Note: this functionality doesn't work right now because
# there is no program gmm-decode-faster-parallel
write_alignments=false
write_words=true
# End configuration section.
@ -59,8 +55,6 @@ if [ $# != 3 ]; then
echo " --cmd (utils/run.pl|utils/queue.pl <queue opts>) # how to run jobs."
echo " --transform-dir <trans-dir> # dir to find fMLLR transforms "
echo " --acwt <float> # acoustic scale used for lattice generation "
#echo " --num-threads <n> # number of threads to use, default 1."
#echo " --parallel-opts <opts> # e.g. '-pe smp 4' if you supply --num-threads 4"
exit 1;
fi
@ -89,8 +83,6 @@ echo "decode.sh: feature type is $feat_type";
splice_opts=`cat $srcdir/splice_opts 2>/dev/null`
cmvn_opts=`cat $srcdir/cmvn_opts 2>/dev/null`
thread_string=
[ $num_threads -gt 1 ] && thread_string="-parallel --num-threads=$num_threads"
case $feat_type in
delta) feats="ark,s,cs:apply-cmvn $cmvn_opts --utt2spk=ark:$sdata/JOB/utt2spk scp:$sdata/JOB/cmvn.scp scp:$sdata/JOB/feats.scp ark:- | add-deltas ark:- ark:- |";;
@ -124,8 +116,8 @@ if [ $stage -le 0 ]; then
[ "`cat $graphdir/num_pdfs`" -eq `am-info --print-args=false $model | grep pdfs | awk '{print $NF}'` ] || \
{ echo "Mismatch in number of pdfs with $model"; exit 1; }
fi
$cmd $parallel_opts JOB=1:$nj $dir/log/decode.JOB.log \
gmm-decode-faster$thread_string --max-active=$max_active --beam=$beam \
$cmd JOB=1:$nj $dir/log/decode.JOB.log \
gmm-decode-faster --max-active=$max_active --beam=$beam \
--acoustic-scale=$acwt --allow-partial=true --word-symbol-table=$graphdir/words.txt \
"$model" $graphdir/HCLG.fst "$feats" "$words" "$ali" || exit 1;
fi

Просмотреть файл

@ -45,7 +45,7 @@ silence_weight=0.01
cmd=run.pl
si_dir=
num_threads=1 # if >1, will use gmm-latgen-faster-parallel
parallel_opts= # If you supply num-threads, you should supply this too.
parallel_opts= # ignored now.
skip_scoring=false
scoring_opts=
# End configuration section
@ -70,7 +70,6 @@ if [ $# != 3 ]; then
echo " # Caution-- must be with same tree"
echo " --acwt <acoustic-weight> # default 0.08333 ... used to get posteriors"
echo " --num-threads <n> # number of threads to use, default 1."
echo " --parallel-opts <opts> # e.g. '-pe smp 4' if you supply --num-threads 4"
echo " --scoring-opts <opts> # options to local/score.sh"
exit 1;
fi
@ -115,7 +114,7 @@ fi
if [ -z "$si_dir" ]; then # we need to do the speaker-independent decoding pass.
si_dir=${dir}.si # Name it as our decoding dir, but with suffix ".si".
if [ $stage -le 0 ]; then
steps/decode.sh --parallel-opts "$parallel_opts" --scoring-opts "$scoring_opts" \
steps/decode.sh --scoring-opts "$scoring_opts" \
--num-threads $num_threads --skip-scoring $skip_scoring \
--acwt $acwt --nj $nj --cmd "$cmd" --beam $first_beam \
--model $alignment_model --max-active \
@ -166,7 +165,7 @@ pass1feats="$pass1splicedfeats transform-feats $srcdir/final.mat ark:- ark:- |"
## model, and it's more correct to store the full state-level lattice for this purpose.
if [ $stage -le 2 ]; then
echo "$0: doing main lattice generation phase"
$cmd $parallel_opts JOB=1:$nj $dir/log/decode.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/decode.JOB.log \
gmm-latgen-faster$thread_string --max-active=$max_active --beam=$beam --lattice-beam=$lattice_beam \
--acoustic-scale=$acwt --determinize-lattice=false \
--allow-partial=true --word-symbol-table=$graphdir/words.txt \
@ -217,7 +216,7 @@ fi
if [ $stage -le 5 ]; then
echo "$0: doing a final pass of acoustic rescoring."
$cmd $parallel_opts JOB=1:$nj $dir/log/acoustic_rescore.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/acoustic_rescore.JOB.log \
gmm-rescore-lattice $final_model "ark:gunzip -c $dir/lat.tmp.JOB.gz|" "$feats" ark:- \| \
lattice-determinize-pruned$thread_string --acoustic-scale=$acwt --beam=$lattice_beam ark:- \
"ark:|gzip -c > $dir/lat.JOB.gz" '&&' rm $dir/lat.tmp.JOB.gz || exit 1;

Просмотреть файл

@ -31,7 +31,7 @@ use_fmllr=false
fmllr_iters=10
fmllr_min_count=1000
num_threads=1 # if >1, will use gmm-latgen-faster-parallel
parallel_opts= # If you supply num-threads, you should supply this too.
parallel_opts= # ignored now.
skip_scoring=false
scoring_opts=
# note: there are no more min-lmwt and max-lmwt options, instead use
@ -132,7 +132,7 @@ if [ $stage -le 2 ]; then
[ "`cat $graphdir/num_pdfs`" -eq `am-info --print-args=false $alignment_model | grep pdfs | awk '{print $NF}'` ] || \
{ echo "Mismatch in number of pdfs with $alignment_model"; exit 1; }
fi
$cmd $parallel_opts JOB=1:$nj $dir/log/decode_pass1.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/decode_pass1.JOB.log \
sgmm2-latgen-faster$thread_string --max-active=$max_active --beam=$beam --lattice-beam=$lattice_beam \
--acoustic-scale=$acwt --determinize-lattice=false --allow-partial=true \
--word-symbol-table=$graphdir/words.txt --max-mem=$max_mem "$gselect_opt_1stpass" $alignment_model \
@ -202,7 +202,7 @@ fi
# corresponding model. Prune and determinize the lattices to limit
# their size.
if [ $stage -le 6 ]; then
$cmd $parallel_opts JOB=1:$nj $dir/log/rescore.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/rescore.JOB.log \
sgmm2-rescore-lattice "$gselect_opt" --utt2spk=ark:$sdata/JOB/utt2spk --spk-vecs=ark:$dir/vecs.JOB \
$srcdir/final.mdl "ark:gunzip -c $dir/pre_lat.JOB.gz|" "$feats" ark:- \| \
lattice-determinize-pruned$thread_string --acoustic-scale=$acwt --beam=$lattice_beam ark:- \

Просмотреть файл

@ -21,7 +21,6 @@ acwt=0.1 # GMM:0.0833, note: only really affects pruning (scoring is on lattices
scoring_opts="--min-lmwt 1 --max-lmwt 12"
skip_scoring=false
use_gpu="no" # disable gpu
parallel_opts=""
# End configuration section.
echo "$0 $@" # Print the command line for logging
@ -104,7 +103,7 @@ fi
# Run the decoding in the queue
if [ $stage -le 0 ]; then
$cmd $parallel_opts JOB=1:$nj $dir/log/decode.JOB.log \
$cmd JOB=1:$nj $dir/log/decode.JOB.log \
nnet-forward --feature-transform=$feature_transform --use-gpu=$use_gpu $nnet "$feats" ark:- \| \
latgen-faster-mapped --max-active=$max_active --max-mem=$max_mem --beam=$beam --lattice-beam=$lattice_beam \
--acoustic-scale=$acwt --allow-partial=true --word-symbol-table=$graphdir/words.txt \

Просмотреть файл

@ -17,7 +17,7 @@ max_mem=20000000 # This will stop the processes getting too large.
# This is in bytes, but not "real" bytes-- you have to multiply
# by something like 5 or 10 to get real bytes (not sure why so large)
num_threads=1
parallel_opts=
parallel_opts= # ignored now
# End configuration section.
echo "$0 $@" # Print the command line for logging
@ -40,7 +40,6 @@ if [ $# != 4 ]; then
echo " # will (individually) finish reasonably soon."
echo " --transform-dir <transform-dir> # directory to find fMLLR transforms."
echo " --num-threads <n> # number of threads per decoding job"
echo " --parallel-opts <string> # if >1 thread, add this to 'cmd', e.g. -pe smp 6"
exit 1;
fi
@ -121,7 +120,7 @@ trap "cleanup" INT QUIT TERM EXIT
if [ $sub_split -eq 1 ]; then
$cmd $parallel_opts JOB=1:$nj $dir/log/decode_den.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/decode_den.JOB.log \
gmm-latgen-faster$thread_string --beam=$beam --lattice-beam=$lattice_beam --acoustic-scale=$acwt \
--max-mem=$max_mem --max-active=$max_active --word-symbol-table=$lang/words.txt $srcdir/final.mdl \
$dir/dengraph/HCLG.fst "$feats" "ark:|gzip -c >$dir/lat.JOB.gz" || exit 1;
@ -147,7 +146,7 @@ else
mkdir -p $dir/part
feats_subset=`echo $feats | sed "s/trans.JOB/trans.$n/g" | sed s:JOB/:$n/split$sub_split/JOB/:g`
$cmd $parallel_opts JOB=1:$sub_split $dir/log/$n/decode_den.JOB.log \
$cmd --num-threads $num_threads JOB=1:$sub_split $dir/log/$n/decode_den.JOB.log \
gmm-latgen-faster$thread_string --beam=$beam --lattice-beam=$lattice_beam --acoustic-scale=$acwt \
--max-mem=$max_mem --max-active=$max_active --word-symbol-table=$lang/words.txt $srcdir/final.mdl \
$dir/dengraph/HCLG.fst "$feats_subset" "ark:|gzip -c >$dir/lat.$n.JOB.gz" || touch $dir/.error &

Просмотреть файл

@ -19,7 +19,7 @@ max_active=5000
transform_dir=
max_mem=20000000 # This will stop the processes getting too large.
num_threads=1
parallel_opts=
parallel_opts= # ignored now.
# End configuration section.
echo "$0 $@" # Print the command line for logging
@ -42,7 +42,6 @@ if [ $# != 4 ]; then
echo " # will (individually) finish reasonably soon."
echo " --transform-dir <transform-dir> # directory to find fMLLR transforms."
echo " --num-threads <n> # number of threads per decoding job"
echo " --parallel-opts <string> # if >1 thread, add this to 'cmd', e.g. -pe smp 6"
exit 1;
fi
@ -57,9 +56,6 @@ cmvn_opts=`cat $srcdir/cmvn_opts 2>/dev/null`
if [ $num_threads -gt 1 ]; then
# the -parallel becomes part of the binary name we decode with.
thread_string="-parallel --num-threads=$num_threads"
if [ -z $parallel_opts ]; then
parallel_opts="--num-threads $num_threads"
fi
fi
mkdir -p $dir/log
@ -142,7 +138,7 @@ cleanup() {
trap "cleanup" INT QUIT TERM EXIT
if [ $sub_split -eq 1 ]; then
$cmd $parallel_opts JOB=1:$nj $dir/log/decode_den.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/decode_den.JOB.log \
sgmm2-latgen-faster$thread_string $spkvecs_opt "$gselect_opt" --beam=$beam \
--lattice-beam=$lattice_beam --acoustic-scale=$acwt \
--max-mem=$max_mem --max-active=$max_active \
@ -171,7 +167,7 @@ else
feats_subset=`echo $feats | sed "s/trans.JOB/trans.$n/g" | sed s:JOB/:$n/split$sub_split/JOB/:g`
spkvecs_opt_subset=`echo $spkvecs_opt | sed "s/JOB/$n/g"`
gselect_opt_subset=`echo $gselect_opt | sed "s/JOB/$n/g"`
$cmd $parallel_opts JOB=1:$sub_split $dir/log/$n/decode_den.JOB.log \
$cmd --num-threads $num_threads JOB=1:$sub_split $dir/log/$n/decode_den.JOB.log \
sgmm2-latgen-faster$thread_string \
$spkvecs_opt_subset "$gselect_opt_subset" \
--beam=$beam --lattice-beam=$lattice_beam \

Просмотреть файл

@ -26,7 +26,7 @@ skip_scoring=false
scoring_opts="--min-lmwt 4 --max-lmwt 15"
num_threads=1 # if >1, will use latgen-faster-parallel
parallel_opts="-pe smp $((num_threads+1))" # use 2 CPUs (1 DNN-forward, 1 decoder)
parallel_opts= # Ignored now.
use_gpu="no" # yes|no|optionaly
# End configuration section.
@ -111,7 +111,7 @@ feats="ark,s,cs:copy-feats scp:$sdata/JOB/feats.scp ark:- |"
# Run the decoding in the queue,
if [ $stage -le 0 ]; then
$cmd $parallel_opts JOB=1:$nj $dir/log/decode.JOB.log \
$cmd --num-threads $((num_threads+1)) JOB=1:$nj $dir/log/decode.JOB.log \
nnet-forward $nnet_forward_opts --feature-transform=$feature_transform --class-frame-counts=$class_frame_counts --use-gpu=$use_gpu $nnet "$feats" ark:- \| \
latgen-faster-mapped$thread_string --min-active=$min_active --max-active=$max_active --max-mem=$max_mem --beam=$beam \
--lattice-beam=$lattice_beam --acoustic-scale=$acwt --allow-partial=true --word-symbol-table=$graphdir/words.txt \

Просмотреть файл

@ -21,7 +21,7 @@ max_mem=20000000 # This will stop the processes getting too large.
# by something like 5 or 10 to get real bytes (not sure why so large)
# End configuration section.
use_gpu=no # yes|no|optional
parallel_opts="-pe smp 2"
parallel_opts="--num-threads 2"
echo "$0 $@" # Print the command line for logging

Просмотреть файл

@ -20,7 +20,7 @@ ivector_scale=1.0
lattice_beam=8.0 # Beam we use in lattice generation.
iter=final
num_threads=1 # if >1, will use gmm-latgen-faster-parallel
parallel_opts= # If you supply num-threads, you should supply this too.
parallel_opts= # ignored now.
scoring_opts=
skip_scoring=false
feat_type=
@ -47,7 +47,7 @@ if [ $# -ne 3 ]; then
echo " --iter <iter> # Iteration of model to decode; default is final."
echo " --scoring-opts <string> # options to local/score.sh"
echo " --num-threads <n> # number of threads to use, default 1."
echo " --parallel-opts <opts> # e.g. '-pe smp 4' if you supply --num-threads 4"
echo " --parallel-opts <opts> # e.g. '--num-threads 4' if you supply --num-threads 4"
exit 1;
fi
@ -129,7 +129,7 @@ if [ ! -z "$online_ivector_dir" ]; then
fi
if [ $stage -le 1 ]; then
$cmd $parallel_opts JOB=1:$nj $dir/log/decode.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/decode.JOB.log \
nnet-latgen-faster$thread_string \
--minimize=$minimize --max-active=$max_active --min-active=$min_active --beam=$beam \
--lattice-beam=$lattice_beam --acoustic-scale=$acwt --allow-partial=true \

Просмотреть файл

@ -19,7 +19,7 @@ max_mem=20000000 # This will stop the processes getting too large.
# by something like 5 or 10 to get real bytes (not sure why so large)
num_threads=1
online_ivector_dir=
parallel_opts=
parallel_opts= # ignored now
feat_type= # you can set this in order to run on top of delta features, although we don't
# normally want to do this.
# End configuration section.
@ -45,7 +45,6 @@ if [ $# != 4 ]; then
echo " # will (individually) finish reasonably soon."
echo " --transform-dir <transform-dir> # directory to find fMLLR transforms."
echo " --num-threads <n> # number of threads per decoding job"
echo " --parallel-opts <string> # if >1 thread, add this to 'cmd', e.g. -pe smp 6"
exit 1;
fi
@ -156,7 +155,7 @@ trap "cleanup" INT QUIT TERM EXIT
if [ $sub_split -eq 1 ]; then
$cmd $parallel_opts JOB=1:$nj $dir/log/decode_den.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/decode_den.JOB.log \
nnet-latgen-faster$thread_string --beam=$beam --lattice-beam=$lattice_beam --acoustic-scale=$acwt \
--max-mem=$max_mem --max-active=$max_active --word-symbol-table=$lang/words.txt $srcdir/final.mdl \
$dir/dengraph/HCLG.fst "$feats" "ark:|gzip -c >$dir/lat.JOB.gz" || exit 1;
@ -182,7 +181,7 @@ else
mkdir -p $dir/part
feats_subset=`echo $feats | sed "s/trans.JOB/trans.$n/g" | sed s:JOB/:$n/split$sub_split/JOB/:g`
$cmd $parallel_opts JOB=1:$sub_split $dir/log/$n/decode_den.JOB.log \
$cmd --num-threads $num_threads JOB=1:$sub_split $dir/log/$n/decode_den.JOB.log \
nnet-latgen-faster$thread_string --beam=$beam --lattice-beam=$lattice_beam --acoustic-scale=$acwt \
--max-mem=$max_mem --max-active=$max_active --word-symbol-table=$lang/words.txt $srcdir/final.mdl \
$dir/dengraph/HCLG.fst "$feats_subset" "ark:|gzip -c >$dir/lat.$n.JOB.gz" || touch $dir/.error &

Просмотреть файл

@ -58,10 +58,10 @@ target_multiplier=0 # Set this to e.g. 1.0 to enable perturbed training.
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G" # by default we use 16 threads; this lets the queue know.
parallel_opts="--num-threads 16 --mem 1G" # by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
combine_num_threads=8
combine_parallel_opts="-pe smp 8" # queue options for the "combine" stage.
combine_parallel_opts="--num-threads 8" # queue options for the "combine" stage.
cleanup=true
egs_dir=
egs_opts=
@ -100,9 +100,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -59,11 +59,11 @@ max_change_per_sample=0.075
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G"
parallel_opts="--num-threads 16 --mem 1G"
# by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
combine_num_threads=8
combine_parallel_opts="-pe smp 8" # queue options for the "combine" stage.
combine_parallel_opts="--num-threads 8" # queue options for the "combine" stage.
cleanup=true
egs_dir=
egs_opts=
@ -101,9 +101,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -43,7 +43,7 @@ widen=0 # If specified, it will increase the hidden-layer dimension
bias_stddev=0.5 # will be used for widen
num_threads=16
parallel_opts="-pe smp $num_threads" # using a smallish #threads by default, out of stability concerns.
parallel_opts="--num-threads $num_threads" # using a smallish #threads by default, out of stability concerns.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
cleanup=true
# End configuration section.
@ -75,7 +75,7 @@ if [ $# != 3 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16\"> # extra options to pass to e.g. queue.pl for processes that"
echo " --parallel-opts <opts|\"--num-threads 16\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -61,7 +61,7 @@ max_change=10.0
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G" # by default we use 16 threads; this lets the queue know.
parallel_opts="--num-threads 16 --mem 1G" # by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
cleanup=true
egs_dir=
@ -104,9 +104,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -37,7 +37,7 @@ io_opts="-tc 5" # for jobs with a lot of I/O, limits the number running at one t
num_threads=16 # this is the default but you may want to change it, e.g. to 1 if
# using GPUs.
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G" # by default we use 4 threads; this lets the queue know.
parallel_opts="--num-threads 16 --mem 1G" # by default we use 4 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
transform_dir= # If this is a SAT system, directory for transforms
cleanup=true
@ -72,9 +72,8 @@ if [ $# != 6 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --samples-per-iter <#samples|400000> # Number of samples of data to process per iteration, per"
echo " # process."

Просмотреть файл

@ -79,8 +79,7 @@ if [ $# != 2 ]; then
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size. With GPU, must be 1."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " # use multiple threads... "
echo " --stage <stage|-3> # Used to run a partially-completed training process from somewhere in"
echo " # the middle."
echo " --criterion <criterion|smbr> # Training criterion: may be smbr, mmi or mpfe"

Просмотреть файл

@ -70,9 +70,8 @@ if [ $# -lt 3 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size. With GPU, must be 1."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --stage <stage|-3> # Used to run a partially-completed training process from somewhere in"
echo " # the middle."
echo " --criterion <criterion|smbr> # Training criterion: may be smbr, mmi or mpfe"

Просмотреть файл

@ -39,7 +39,7 @@ shuffle_buffer_size=5000 # This "buffer_size" variable controls randomization of
mix_up=0
stage=-5
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G" # by default we use 16 threads; this lets the queue know.
parallel_opts="--num-threads 16 --mem 1G" # by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
cleanup=true
remove_egs=false
@ -69,9 +69,8 @@ if [ $# != 3 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."
echo " --num-iters-final <#iters|20> # Number of final iterations to give to nnet-combine-fast to "

Просмотреть файл

@ -43,7 +43,7 @@ num_jobs_nnet=4
mix_up=0
stage=-5
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G" # by default we use 16 threads; this lets the queue know.
parallel_opts="--num-threads 16 --mem 1G" # by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
combine_num_threads=8
cleanup=true
@ -76,9 +76,8 @@ if [ $# != 3 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."
echo " --num-iters-final <#iters|20> # Number of final iterations to give to nnet-combine-fast to "

Просмотреть файл

@ -111,9 +111,8 @@ if [ $# -lt 6 -o $[$#%2] -ne 0 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --stage <stage|-4> # Used to run a partially-completed training process from somewhere in"
echo " # the middle."
exit 1;

Просмотреть файл

@ -76,11 +76,11 @@ precondition_rank_out=80 # relates to online preconditioning
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G"
parallel_opts="--num-threads 16 --mem 1G"
# by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
combine_num_threads=8
combine_parallel_opts="-pe smp 8" # queue options for the "combine" stage.
combine_parallel_opts="--num-threads 8" # queue options for the "combine" stage.
cleanup=true
egs_dir=
lda_opts=
@ -131,9 +131,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -74,11 +74,11 @@ precondition_rank_out=80 # relates to online preconditioning
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G"
parallel_opts="--num-threads 16 --mem 1G"
# by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
combine_num_threads=8
combine_parallel_opts="-pe smp 8" # queue options for the "combine" stage.
combine_parallel_opts="--num-threads 8" # queue options for the "combine" stage.
cleanup=true
egs_dir=
lda_opts=
@ -129,9 +129,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -64,7 +64,7 @@ max_change=10.0
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G" # by default we use 16 threads; this lets the queue know.
parallel_opts="--num-threads 16 --mem 1G" # by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
cleanup=true
egs_dir=
@ -116,9 +116,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -77,11 +77,11 @@ precondition_rank_out=80 # relates to online preconditioning
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G"
parallel_opts="--num-threads 16 --mem 1G"
# by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
combine_num_threads=8
combine_parallel_opts="-pe smp 8" # queue options for the "combine" stage.
combine_parallel_opts="--num-threads 8" # queue options for the "combine" stage.
cleanup=true
egs_dir=
lda_opts=
@ -129,9 +129,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -70,10 +70,10 @@ target_multiplier=0 # Set this to e.g. 1.0 to enable perturbed training.
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G" # by default we use 16 threads; this lets the queue know.
parallel_opts="--num-threads 16 --mem 1G" # by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
combine_num_threads=8
combine_parallel_opts="-pe smp 8" # queue options for the "combine" stage.
combine_parallel_opts="--num-threads 8" # queue options for the "combine" stage.
cleanup=true
egs_dir=
lda_opts=
@ -122,9 +122,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -62,7 +62,7 @@ max_change=10.0
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G" # by default we use 16 threads; this lets the queue know.
parallel_opts="--num-threads 16 --mem 1G" # by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
cleanup=true
egs_dir=
@ -107,9 +107,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -72,10 +72,10 @@ target_multiplier=0 # Set this to e.g. 1.0 to enable perturbed training.
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G" # by default we use 16 threads; this lets the queue know.
parallel_opts="--num-threads 16 --mem 1G" # by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
combine_num_threads=8
combine_parallel_opts="-pe smp 8" # queue options for the "combine" stage.
combine_parallel_opts="--num-threads 8" # queue options for the "combine" stage.
cleanup=true
egs_dir=
lda_opts=
@ -124,9 +124,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -77,11 +77,11 @@ precondition_rank_out=80 # relates to online preconditioning
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G"
parallel_opts="--num-threads 16 --mem 1G"
# by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
combine_num_threads=8
combine_parallel_opts="-pe smp 8" # queue options for the "combine" stage.
combine_parallel_opts="--num-threads 8" # queue options for the "combine" stage.
cleanup=true
egs_dir=
lda_opts=
@ -128,9 +128,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -72,11 +72,11 @@ precondition_rank_out=80 # relates to online preconditioning
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G"
parallel_opts="--num-threads 16 --mem 1G"
# by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
combine_num_threads=8
combine_parallel_opts="-pe smp 8" # queue options for the "combine" stage.
combine_parallel_opts="--num-threads 8" # queue options for the "combine" stage.
cleanup=true
egs_dir=
lda_opts=
@ -121,9 +121,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -74,11 +74,11 @@ precondition_rank_out=80 # relates to online preconditioning
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G"
parallel_opts="--num-threads 16 --mem 1G"
# by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
combine_num_threads=8
combine_parallel_opts="-pe smp 8" # queue options for the "combine" stage.
combine_parallel_opts="--num-threads 8" # queue options for the "combine" stage.
cleanup=true
egs_dir=
lda_opts=
@ -125,9 +125,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -83,11 +83,11 @@ precondition_rank_out=80 # relates to online preconditioning
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G"
parallel_opts="--num-threads 16 --mem 1G"
# by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
combine_num_threads=8
combine_parallel_opts="-pe smp 8" # queue options for the "combine" stage.
combine_parallel_opts="--num-threads 8" # queue options for the "combine" stage.
cleanup=true
egs_dir=
lda_opts=
@ -135,9 +135,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -59,7 +59,7 @@ max_change=10.0
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G" # by default we use 16 threads; this lets the queue know.
parallel_opts="--num-threads 16 --mem 1G" # by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
cleanup=true
egs_dir=
@ -108,9 +108,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -63,7 +63,7 @@ max_change=10.0
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G" # by default we use 16 threads; this lets the queue know.
parallel_opts="--num-threads 16 --mem 1G" # by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
cleanup=true
egs_dir=
@ -108,9 +108,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -71,9 +71,9 @@ precondition_rank_out=60 # relates to online preconditioning
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G" # by default we use 16 threads; this lets the queue know.
parallel_opts="--num-threads 16 --mem 1G" # by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
combine_parallel_opts="-pe smp 8" # queue options for the "combine" stage.
combine_parallel_opts="--num-threads 8" # queue options for the "combine" stage.
combine_num_threads=8
cleanup=true
egs_dir=
@ -122,9 +122,8 @@ if [ $# != 4 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -48,7 +48,7 @@ max_change=10.0
mix_up=0 # Number of components to mix up to (should be > #tree leaves, if
# specified.)
num_threads=16
parallel_opts="-pe smp 16 -l ram_free=1G,mem_free=1G" # by default we use 16 threads; this lets the queue know.
parallel_opts="--num-threads 16 --mem 1G" # by default we use 16 threads; this lets the queue know.
# note: parallel_opts doesn't automatically get adjusted if you adjust num-threads.
cleanup=false
egs_dir=
@ -81,9 +81,8 @@ if [ $# != 5 ]; then
echo " --num-threads <num-threads|16> # Number of parallel threads per job (will affect results"
echo " # as well as speed; may interact with batch size; if you increase"
echo " # this, you may want to decrease the batch size."
echo " --parallel-opts <opts|\"-pe smp 16 -l ram_free=1G,mem_free=1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... note, you might have to reduce mem_free,ram_free"
echo " # versus your defaults, because it gets multiplied by the -pe smp argument."
echo " --parallel-opts <opts|\"--num-threads 16 --mem 1G\"> # extra options to pass to e.g. queue.pl for processes that"
echo " # use multiple threads... "
echo " --io-opts <opts|\"-tc 10\"> # Options given to e.g. queue.pl for jobs that do a lot of I/O."
echo " --minibatch-size <minibatch-size|128> # Size of minibatch to process (note: product with --num-threads"
echo " # should not get too large, e.g. >2k)."

Просмотреть файл

@ -19,7 +19,7 @@ max_mem=20000000 # This will stop the processes getting too large.
# This is in bytes, but not "real" bytes-- you have to multiply
# by something like 5 or 10 to get real bytes (not sure why so large)
num_threads=1
parallel_opts=
parallel_opts= # ignored now.
# End configuration section.
echo "$0 $@" # Print the command line for logging
@ -41,7 +41,6 @@ if [ $# != 4 ]; then
echo " # large databases so your jobs will be smaller and"
echo " # will (individually) finish reasonably soon."
echo " --num-threads <n> # number of threads per decoding job"
echo " --parallel-opts <string> # if >1 thread, add this to 'cmd', e.g. -pe smp 6"
exit 1;
fi
@ -115,7 +114,7 @@ trap "cleanup" INT QUIT TERM EXIT
if [ $sub_split -eq 1 ]; then
$cmd $parallel_opts JOB=1:$nj $dir/log/decode_den.JOB.log \
$cmd --num-threads $num_threads JOB=1:$nj $dir/log/decode_den.JOB.log \
nnet-latgen-faster$thread_string --beam=$beam --lattice-beam=$lattice_beam --acoustic-scale=$acwt \
--max-mem=$max_mem --max-active=$max_active --word-symbol-table=$lang/words.txt $srcdir/final.mdl \
$dir/dengraph/HCLG.fst "$feats" "ark:|gzip -c >$dir/lat.JOB.gz" || exit 1;
@ -141,7 +140,7 @@ else
mkdir -p $dir/part
feats_subset=`echo $feats | sed "s/trans.JOB/trans.$n/g" | sed s:JOB/:$n/split$sub_split/JOB/:g`
$cmd $parallel_opts JOB=1:$sub_split $dir/log/$n/decode_den.JOB.log \
$cmd --num-threads $num_threads JOB=1:$sub_split $dir/log/$n/decode_den.JOB.log \
nnet-latgen-faster$thread_string --beam=$beam --lattice-beam=$lattice_beam --acoustic-scale=$acwt \
--max-mem=$max_mem --max-active=$max_active --word-symbol-table=$lang/words.txt $srcdir/final.mdl \
$dir/dengraph/HCLG.fst "$feats_subset" "ark:|gzip -c >$dir/lat.$n.JOB.gz" || touch $dir/.error &

Просмотреть файл

@ -36,7 +36,7 @@ cleanup=true
min_gaussian_weight=0.0001
remove_low_count_gaussians=true # set this to false if you need #gauss to stay fixed.
num_threads=32
parallel_opts="-pe smp 32"
parallel_opts= # ignored now.
online_cmvn_config=conf/online_cmvn.conf
# End configuration section.
@ -68,8 +68,6 @@ if [ $# != 4 ]; then
echo " # in initialization phase (then split)"
echo " --num-threads <n|32> # number of threads to use in initialization"
echo " # phase (must match with parallel-opts option)"
echo " --parallel-opts <string|'-pe smp 32'> # Option should match number of threads in"
echo " # --num-threads option above"
echo " --min-gaussian-weight <weight|0.0001> # min Gaussian weight allowed in GMM"
echo " # initialization (this relatively high"
echo " # value keeps counts fairly even)"
@ -115,7 +113,7 @@ if [ $stage -le -2 ]; then
echo "$0: starting from $num_gauss_init Gaussians, reaching $num_gauss;"
echo "$0: for $num_iters_init iterations, using at most $num_frames frames of data"
$cmd $parallel_opts $dir/log/gmm_init.log \
$cmd --num-threads $num_threads $dir/log/gmm_init.log \
gmm-global-init-from-feats --num-threads=$num_threads --num-frames=$num_frames \
--min-gaussian-weight=$min_gaussian_weight \
--num-gauss=$num_gauss --num-gauss-init=$num_gauss_init --num-iters=$num_iters_init \

Просмотреть файл

@ -51,7 +51,7 @@ min_post=0.025 # Minimum posterior to use (posteriors below this are pruned out)
subsample=2 # This speeds up the training: training on every 2nd feature
# (configurable) Since the features are highly correlated across
# frames, we don't expect to lose too much from this.
parallel_opts= #Task running engine configuration
parallel_opts= # ignored now.
cleanup=true
# End configuration section.
@ -76,8 +76,6 @@ if [ $# != 3 ]; then
echo " --stage <stage|-4> # To control partial reruns"
echo " --num-gselect <n|5> # Number of Gaussians to select using"
echo " # diagonal model."
echo " --parallel-opts <opts> # e.g. '-pe smp 16 ', the number should be equivalent"
echo " # to --num-processes * --num-threads"
exit 1;
fi
@ -107,11 +105,6 @@ gmm_feats="ark,s,cs:apply-cmvn-online --config=$dir/online_cmvn.conf $dir/global
feats="ark,s,cs:splice-feats $splice_opts scp:$sdata/JOB/feats.scp ark:- | transform-feats $dir/final.mat ark:- ark:- | subsample-feats --n=$subsample ark:- ark:- |"
#We will specify our own parallel-opts only in cases user does not supply anything.
#If user does specify parallel-opts, then we will assume user knows what's right
if [ -z "$parallel_opts" ] ; then
parallel_opts="-pe smp $[$num_threads*$num_processes]"
fi
# Initialize the i-vector extractor using the input GMM, which is converted to
# full because that's what the i-vector extractor expects. Note: we have to do
@ -157,7 +150,7 @@ while [ $x -lt $num_iters ]; do
echo "Accumulating stats (pass $x)"
for g in $(seq $nj); do
start=$[$num_processes*($g-1)+1]
$cmd $parallel_opts $dir/log/acc.$x.$g.log \
$cmd --num-threads $[$num_threads*$num_processes] $dir/log/acc.$x.$g.log \
ivector-extractor-sum-accs --parallel=true "${Args[@]:$start:$num_processes}" \
$dir/acc.$x.$g || touch $dir/.error &
done
@ -178,7 +171,7 @@ while [ $x -lt $num_iters ]; do
# The parallel-opts was either specified by
# the user or we computed it correctly in
# tge previous stages
$cmd $parallel_opts $dir/log/update.$x.log \
$cmd --num-threads $[$num_threads*$num_processes] $dir/log/update.$x.log \
ivector-extractor-est --num-threads=$nt $dir/$x.ie $dir/acc.$x $dir/$[$x+1].ie || exit 1;
rm $dir/acc.$x.*
if $cleanup; then

Просмотреть файл

@ -80,9 +80,10 @@ if (@ARGV > 0) {
}
}
if ($ignored_opts ne "") {
print STDERR "run.pl: Warning: ignoring options \"$ignored_opts\"\n";
}
# Users found this message confusing so we are removing it.
# if ($ignored_opts ne "") {
# print STDERR "run.pl: Warning: ignoring options \"$ignored_opts\"\n";
# }
if ($max_jobs_run == -1) { # If --max-jobs-run option not set,
# then work out the number of processors if possible,