I try librispeech recipe.
When I run Generating training examples on disk, it fails.
Error is as follows.
steps/nnet3/chain/get_egs.sh: Generating training examples on disk
run.pl: 99 / 100 failed, log is in exp/chain_cleaned/tdnn_1d_sp/egs/log/get_egs.*.log
Traceback (most recent call last):
File "steps/nnet3/chain/train.py", line 637, in main
train(args, run_opts)
File "steps/nnet3/chain/train.py", line 396, in train
stage=args.egs_stage
File "steps/libs/nnet3/train/chain_objf/acoustic_model.py", line 118, in generate_chain_egs
egs_opts=egs_opts if egs_opts is not None else ''))
File "steps/libs/common.py", line 158, in execute_command
p.returncode, command))
Exception: Command exited with status 1: steps/nnet3/chain/get_egs.sh --frames-overlap-per-eg 0 --constrained false --cmd "
run.pl" --cmvn-opts "--norm-means=false --norm-vars=false" --online-ivector-dir "exp/nnet3_cleaned/ivectors_train_clean_100_sp_hires" --left-context 41 --right-context 41 --left-context-initial -1 --right-context-final -1 --left-tolerance '5' --right-tolerance '5' --frame-subsampling-factor 3 --alignment-subsampling-factor 3 --stage -10 --frames-per-iter 2500000 --frames-per-eg 150,110,100 --srand 0 /nas206ce8_homes/takagi/kaldi/egs/librispeech/s5/data/train_clean_100_sp_hires exp/chain_cleaned/tdnn_1d_sp exp/chain_cleaned/tri4b_cleaned_train_clean_100_sp_lats exp/chain_cleaned/tdnn_1d_sp/egs
steps/nnet3/chain/train.py --stage -10 --cmd
run.pl --feat.online-ivector-dir exp/nnet3_cleaned/ivectors_train_clean_100_sp_hires --feat.cmvn-opts --norm-means=false --norm-vars=false --chain.xent-regularize 0.1 --chain.leaky-hmm-coefficient 0.1 --chain.l2-regularize 0.0 --chain.apply-deriv-weights false --chain.lm-opts=--num-extra-lm-states=2000 --egs.dir --egs.stage -10 --egs.opts --frames-overlap-per-eg 0 --constrained false --egs.chunk-width 150,110,100 --trainer.dropout-schedule 0,0...@0.20,0...@0.50,0 --trainer.add-option=--optimization.memory-compression-level=2 --trainer.num-chunk-per-minibatch 64 --trainer.frames-per-iter 2500000 --trainer.num-epochs 4 --trainer.optimization.num-jobs-initial 3 --trainer.optimization.num-jobs-final 16 --trainer.optimization.initial-effective-lrate 0.00015 --trainer.optimization.final-effective-lrate 0.000015 --trainer.max-param-change 2.0 --cleanup.remove-egs true --feat-dir /nas206ce8_homes/takagi/kaldi/egs/librispeech/s5/data/train_clean_100_sp_hires --tree-dir exp/chain_cleaned/tree_sp --lat-dir exp/chain_cleaned/tri4b_cleaned_train_clean_100_sp_lats --dir exp/chain_cleaned/tdnn_1d_sp
['steps/nnet3/chain/train.py', '--stage', '-10', '--cmd', '
run.pl', '--feat.online-ivector-dir', 'exp/nnet3_cleaned/ivectors_train_clean_100_sp_hires', '--feat.cmvn-opts', '--norm-means=false --norm-vars=false', '--chain.xent-regularize', '0.1', '--chain.leaky-hmm-coefficient', '0.1', '--chain.l2-regularize', '0.0', '--chain.apply-deriv-weights', 'false', '--chain.lm-opts=--num-extra-lm-states=2000', '--egs.dir', '', '--egs.stage', '-10', '--egs.opts', '--frames-overlap-per-eg 0 --constrained false', '--egs.chunk-width', '150,110,100', '--trainer.dropout-schedule', '0,0...@0.20,0...@0.50,0', '--trainer.add-option=--optimization.memory-compression-level=2', '--trainer.num-chunk-per-minibatch', '64', '--trainer.frames-per-iter', '2500000', '--trainer.num-epochs', '4', '--trainer.optimization.num-jobs-initial', '3', '--trainer.optimization.num-jobs-final', '16', '--trainer.optimization.initial-effective-lrate', '0.00015', '--trainer.optimization.final-effective-lrate', '0.000015', '--trainer.max-param-change', '2.0', '--cleanup.remove-egs', 'true', '--feat-dir', '/nas206ce8_homes/takagi/kaldi/egs/librispeech/s5/data/train_clean_100_sp_hires', '--tree-dir', 'exp/chain_cleaned/tree_sp', '--lat-dir', 'exp/chain_cleaned/tri4b_cleaned_train_clean_100_sp_lats', '--dir', 'exp/chain_cleaned/tdnn_1d_sp']
steps/nnet3/chain/get_egs.sh: Getting subsets of validation examples for diagnostics and combination.
# lattice-align-phones --replace-output-symbols=true exp/chain_cleaned/tri4b_cleaned_train_clean_100_sp_lats/final.mdl "ark:gunzip -c exp/chain_cleaned/tri4b_cleaned_train_clean_100_sp_lats/lat.1.gz |" ark:- | chain-get-supervision --lattice-input=true --frame-subsampling-factor=3 --right-tolerance=5 --left-tolerance=5 --convert-to-pdfs=false exp/chain_cleaned/tdnn_1d_sp/tree exp/chain_cleaned/tdnn_1d_sp/0.trans_mdl ark:- ark:- | nnet3-chain-get-egs --online-ivectors=scp:exp/nnet3_cleaned/ivectors_train_clean_100_sp_hires/ivector_online.scp --online-ivector-period=10 --srand=$[1+0] --left-context=41 --right-context=41 --num-frames=150,110,100 --frame-subsampling-factor=3 --compress=true --num-frames-overlap=0 --transition-model=exp/chain_cleaned/tdnn_1d_sp/0.trans_mdl "ark,s,cs:utils/
filter_scp.pl --exclude exp/chain_cleaned/tdnn_1d_sp/egs/valid_uttlist /nas206ce8_homes/takagi/kaldi/egs/librispeech/s5/data/train_clean_100_sp_hires/split100/1/feats.scp | apply-cmvn --norm-means=false --norm-vars=false --utt2spk=ark:/nas206ce8_homes/takagi/kaldi/egs/librispeech/s5/data/train_clean_100_sp_hires/split100/1/utt2spk scp:/nas206ce8_homes/takagi/kaldi/egs/librispeech/s5/data/train_clean_100_sp_hires/split100/1/cmvn.scp scp:- ark:- |" ark,s,cs:- ark:- | nnet3-chain-copy-egs --random=true --srand=$[1+0] ark:- ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.1.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.2.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.3.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.4.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.5.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.6.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.7.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.8.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.9.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.10.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.11.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.12.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.13.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.14.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.15.ark
# Started at Wed Aug 28 20:26:28 JST 2019
#
lattice-align-phones --replace-output-symbols=true exp/chain_cleaned/tri4b_cleaned_train_clean_100_sp_lats/final.mdl 'ark:gunzip -c exp/chain_cleaned/tri4b_cleaned_train_clean_100_sp_lats/lat.1.gz |' ark:-
nnet3-chain-copy-egs --random=true --srand=1 ark:- ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.1.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.2.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.3.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.4.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.5.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.6.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.7.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.8.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.9.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.10.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.11.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.12.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.13.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.14.ark ark:exp/chain_cleaned/tdnn_1d_sp/egs/cegs_orig.1.15.ark
nnet3-chain-get-egs --online-ivectors=scp:exp/nnet3_cleaned/ivectors_train_clean_100_sp_hires/ivector_online.scp --online-ivector-period=10 --srand=1 --left-context=41 --right-context=41 --num-frames=150,110,100 --frame-subsampling-factor=3 --compress=true --num-frames-overlap=0 --transition-model=exp/chain_cleaned/tdnn_1d_sp/0.trans_mdl 'ark,s,cs:utils/
filter_scp.pl --exclude exp/chain_cleaned/tdnn_1d_sp/egs/valid_uttlist /nas206ce8_homes/takagi/kaldi/egs/librispeech/s5/data/train_clean_100_sp_hires/split100/1/feats.scp | apply-cmvn --norm-means=false --norm-vars=false --utt2spk=ark:/nas206ce8_homes/takagi/kaldi/egs/librispeech/s5/data/train_clean_100_sp_hires/split100/1/utt2spk scp:/nas206ce8_homes/takagi/kaldi/egs/librispeech/s5/data/train_clean_100_sp_hires/split100/1/cmvn.scp scp:- ark:- |' ark,s,cs:- ark:-
LOG (nnet3-chain-get-egs[5.5.356~1-8706]:ComputeDerived():nnet-example-utils.cc:335) Rounding up --num-frames=150,110,100 to multiples of --frame-subsampling-factor=3, to: 150,111,102
chain-get-supervision --lattice-input=true --frame-subsampling-factor=3 --right-tolerance=5 --left-tolerance=5 --convert-to-pdfs=false exp/chain_cleaned/tdnn_1d_sp/tree exp/chain_cleaned/tdnn_1d_sp/0.trans_mdl ark:- ark:-
apply-cmvn --norm-means=false --norm-vars=false --utt2spk=ark:/nas206ce8_homes/takagi/kaldi/egs/librispeech/s5/data/train_clean_100_sp_hires/split100/1/utt2spk scp:/nas206ce8_homes/takagi/kaldi/egs/librispeech/s5/data/train_clean_100_sp_hires/split100/1/cmvn.scp scp:- ark:-
LOG (apply-cmvn[5.5.356~1-8706]:main():apply-cmvn.cc:81) Copied 277 utterances.
LOG (nnet3-chain-get-egs[5.5.356~1-8706]:~UtteranceSplitter():nnet-example-utils.cc:357) Split 277 utts, with total length 334071 frames (0.927975 hours assuming 100 frames per second)
LOG (nnet3-chain-get-egs[5.5.356~1-8706]:~UtteranceSplitter():nnet-example-utils.cc:366) Average chunk length was 144.227 frames; overlap between adjacent chunks was 0.609751% of input length; length of output was 100.462% of input length (minus overlap = 99.8527%).
LOG (nnet3-chain-get-egs[5.5.356~1-8706]:~UtteranceSplitter():nnet-example-utils.cc:382) Output frames are distributed among chunk-sizes as follows: 102 = 6.382%, 111 = 2.844%, 150 = 90.77%
ERROR: CompactFst write failed: <unknown>
ERROR (chain-get-supervision[5.5.356~1-8706]:WriteToken():io-funcs.cc:141) Write failure in WriteToken.
[ Stack-Trace: ]
kaldi::MessageLogger::LogMessage() const
kaldi::MessageLogger::LogAndThrow::operator=(kaldi::MessageLogger const&)
kaldi::WriteToken(std::ostream&, bool, char const*)
kaldi::chain::Supervision::Write(std::ostream&, bool) const
kaldi::KaldiObjectHolder<kaldi::chain::Supervision>::Write(std::ostream&, bool, kaldi::chain::Supervision const&)
kaldi::TableWriterArchiveImpl<kaldi::KaldiObjectHolder<kaldi::chain::Supervision> >::Write(std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> > const&, kaldi::chain::Supervision const&)
chain-get-supervision() [0x546df1]
main
__libc_start_main
_start
WARNING (chain-get-supervision[5.5.356~1-8706]:Write():util/kaldi-holder-inl.h:57) Exception caught writing Table object. kaldi::KaldiFatalError
WARNING (chain-get-supervision[5.5.356~1-8706]:Write():util/kaldi-table-inl.h:1057) Write failure to standard output
ERROR (chain-get-supervision[5.5.356~1-8706]:Write():util/kaldi-table-inl.h:1515) Error in TableWriter::Write
[ Stack-Trace: ]
kaldi::MessageLogger::LogMessage() const
kaldi::MessageLogger::LogAndThrow::operator=(kaldi::MessageLogger const&)
chain-get-supervision() [0x546e43]
main
__libc_start_main
_start
WARNING (chain-get-supervision[5.5.356~1-8706]:Close():util/kaldi-table-inl.h:1089) Error closing stream: wspecifier is ark:-
ERROR (chain-get-supervision[5.5.356~1-8706]:~TableWriter():util/kaldi-table-inl.h:1539) Error closing TableWriter [in destructor].
[ Stack-Trace: ]
kaldi::MessageLogger::LogMessage() const
kaldi::MessageLogger::LogAndThrow::operator=(kaldi::MessageLogger const&)
kaldi::TableWriter<kaldi::KaldiObjectHolder<kaldi::chain::Supervision> >::~TableWriter()
main
__libc_start_main
_start
terminate called after throwing an instance of 'kaldi::KaldiFatalError'
what(): kaldi::KaldiFatalError
ERROR: VectorFst::Write: Write failed: <unspecifed>
WARNING (lattice-align-phones[5.5.356~1-8706]:Write():util/kaldi-table-inl.h:1057) Write failure to standard output
ERROR (lattice-align-phones[5.5.356~1-8706]:Write():util/kaldi-table-inl.h:1515) Error in TableWriter::Write
[ Stack-Trace: ]
kaldi::MessageLogger::LogMessage() const
kaldi::MessageLogger::LogAndThrow::operator=(kaldi::MessageLogger const&)
main
__libc_start_main
_start
WARNING (lattice-align-phones[5.5.356~1-8706]:Close():util/kaldi-table-inl.h:1089) Error closing stream: wspecifier is ark:-
ERROR (lattice-align-phones[5.5.356~1-8706]:~TableWriter():util/kaldi-table-inl.h:1539) Error closing TableWriter [in destructor].
[ Stack-Trace: ]
kaldi::MessageLogger::LogMessage() const
kaldi::MessageLogger::LogAndThrow::operator=(kaldi::MessageLogger const&)
kaldi::TableWriter<kaldi::CompactLatticeHolder>::~TableWriter()
main
__libc_start_main
_start
terminate called after throwing an instance of 'kaldi::KaldiFatalError'
what(): kaldi::KaldiFatalError
gzip: stdout: Broken pipe
LOG (nnet3-chain-copy-egs[5.5.356~1-8706]:main():nnet3-chain-copy-egs.cc:395) Read 2327 neural-network training examples, wrote 2327
# Accounting: time=11 threads=1
# Ended (code 0) at Wed Aug 28 20:26:39 JST 2019, elapsed time 11 seconds