Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

add aishell recipe #1742

Merged
merged 3 commits into from
Jul 17, 2017
Merged

add aishell recipe #1742

merged 3 commits into from
Jul 17, 2017

Conversation

naxingyu
Copy link
Contributor

@naxingyu naxingyu commented Jul 6, 2017

Add recipe for AIShell corpus, which is recently added to http://www.openslr.org/33/

@danpovey
Copy link
Contributor

danpovey commented Jul 6, 2017 via email

@keli78
Copy link
Contributor

keli78 commented Jul 6, 2017

Sure, will check it soon.

@keli78
Copy link
Contributor

keli78 commented Jul 6, 2017

I can avoid downloading the data while I got two errors to run the run.sh by far:

  1. The first line in path.sh is not correct as I got an error "The standard file ../../tools/config/common_path.sh is not present -> Exit!"
  2. With it fixed, I got another one as below:
    local/aishell_train_lms.sh: line 47: get_word_map.pl: command not found
    I guess the problem is because the way it checks whether kaldi_lm is installed is not correct.
    After I fixed this issue, it runs ok by far.

@naxingyu naxingyu force-pushed the aishell branch 2 times, most recently from 30771fc to e4418cc Compare July 6, 2017 06:51
@naxingyu
Copy link
Contributor Author

naxingyu commented Jul 6, 2017 via email

n=`cat $train_dir/wav.flist $dev_dir/wav.flist $test_dir/wav.flist | wc -l`
[ $n -ne 141925 ] && \
echo Warning: expected 141925 data data files, found $n

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think this is not particularly effective, running the same find three times over the same directory -- how about running it once, getting single list of files, comparing the number of lines and only after that using grep on this files to get the partial filelists?

fsttablecompose data/lang/L_disambig.fst data/lang_test/G.fst | \
fstisstochastic || echo LG is not stochastic

echo "$0: AISHELL data formatting succeeded"
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

there is a script format_lm.sh in the utils dir -- perhaps you could call that one?

cat $dir/word_map | awk '{print $1}' | cat - <(echo "<s>"; echo "</s>" ) > $sdir/wordlist


ngram-count -text $sdir/train -order 3 -limit-vocab -vocab $sdir/wordlist -unk \
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

are you testing srilm is installed?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This is just for showing an example of using srilm. Normal running of this script will exit before the srilm part.

#!/bin/bash

# This script is modified based on swbd/s5c/local/nnet3/run_ivector_common.sh

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think @danpovey was saying the minilibrispeech is a more recent script which should be preferable.

# Train a system just for its LDA+MLLT transform. We use --num-iters 13
# because after we get the transform (12th iter is the last), any further
# training is pointless.
steps/train_lda_mllt.sh --cmd "$train_cmd" --num-iters 13 \
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

can you try replacing this with the pca transform training -- the parameters are the same (if you decide not to go for the minils ivector-common script)

@danpovey
Copy link
Contributor

danpovey commented Jul 6, 2017 via email

@naxingyu
Copy link
Contributor Author

naxingyu commented Jul 12, 2017 via email

@naxingyu
Copy link
Contributor Author

@danpovey @jtrmal Thanks for the suggestions. I've changed the i-Vector script and re-run the related experiments, and results are updated. Please check out.

@danpovey
Copy link
Contributor

@keli78, can you please check if this still runs?

@keli78
Copy link
Contributor

keli78 commented Jul 13, 2017

Sure.

@keli78
Copy link
Contributor

keli78 commented Jul 14, 2017

It's in train mono-phone stage and no problem occurs by far.

--egs.dir "$common_egs_dir" \
--egs.stage $get_egs_stage \
--egs.opts "--frames-overlap-per-eg 0" \
--egs.chunk-width $frames_per_eg \
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I just noticed some very small issues with the nnet3+chain recipe.
Can you please rename this from 7h to 1a, and put it in the tuning/ directory, and create a soft link?
Also can you make frames_per_eg a comma-separated list, like 150,110,90 ?
Also I think you'll get better results if you change renorm to batchnorm.

@naxingyu
Copy link
Contributor Author

@danpovey results updated.

@danpovey danpovey merged commit aedc2fe into kaldi-asr:master Jul 17, 2017
kronos-cm added a commit to kronos-cm/kaldi that referenced this pull request Jul 25, 2017
* 'master' of https://github.com/kaldi-asr/kaldi: (36 commits)
  [scripts] Fix convert_nnet2_to_nnet3.py (kaldi-asr#1774)
  [egs] Add missing make_corpus_subset.sh in babel_multilang example (kaldi-asr#1766)
  [egs] Graphemic lexicon updates / fixes in babel/s5d recipe and hub4_spanish recipe (kaldi-asr#1740)
  [egs] update hkust results (kaldi-asr#1772)
  [egs] Update AMI chain experiments RE dropout, decay-time and proportional-shrink (kaldi-asr#1732)
  [egs] Fixes to the aishell (Mandarin) recipe (kaldi-asr#1770)
  [egs] Add recipe for aishell data (free Mandarin corpus, 170 hours total) (kaldi-asr#1742)
  [src] Change to arpa-reading code to accept blank lines with whitespace (kaldi-asr#1752)
  [scripts] For nnet3 training, add option to disable the model-combination  (kaldi-asr#1757)
  [scripts] minor bugfix to nnet1 alignment script when creating lattices (kaldi-asr#1764)
  [src] Add support for row/column ranges when reading GeneralMatrix (kaldi-asr#1761)
  [src] Change name of option --norm-mean->--norm-means for consistency, thanks: 415198468@qq.com
  [egs] swbd/s5c, added 5 layer (b)lstm recipes (kaldi-asr#1759)
  [scripts] Fix bug in segment_long_utterances.sh (kaldi-asr#1758)
  [src] Fix indexing error in nnet1::Convolutional2DComponent (kaldi-asr#1755)
  [src] Fix usage message of program (thanks:jubang0219@gmail.com)
  [egs] some small updates to scripts (installing beamformit; segmentation example)
  [egs] Small fix to ami/s5b/local/chain/compare_wer_general.sh (kaldi-asr#1751)
  [build] Add configuration check for incompatible g++ compilers when CUDA is enabled. (kaldi-asr#1749)
  [egs] Update Librispeech nnet3 TDNN recipe (old one did not run) (kaldi-asr#1727)
  ...
@naxingyu naxingyu deleted the aishell branch September 16, 2017 15:26
Skaiste pushed a commit to Skaiste/idlak that referenced this pull request Sep 26, 2018
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants