Blame view

egs/voxforge/online_demo/README.txt 1.18 KB
8dcb6dfcb   Yannick Estève   first commit
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
  This directory contains a small demo, which gives an example of how the online
  decoding binaries can be used. These binaries are optional so if you want to use
  them please type "make ext" in KALDI_ROOT/src. Pre-built acoustic models are 
  utilized, which were trained on VoxForge's data(using egs/voxforge/s5). The 
  decoding graphs are constructed using a language model built on a public domain 
  book's text. There are two test modes: simulated online decoding, where the audio
  comes from pre-recorded audio snippets, and live decoding mode, where the user can 
  try how his speech is recognized with the afore-mentioned language and acoustic
  models. The microphone input is implemented using PortAudio library.
  The audio(for the "simulated input" mode) and the models are automatically
  downloaded the first time when the script is started.
  
  You can start the demo in "simulated" mode by just calling:
  ./run.sh
  
  To try using your own voice(assuming you have a working microphone) call:
  ./run.sh --test-mode live
  
  By default a discriminatively trained acoustic models is used. To try a
  maximum likelihood trained one(usually has lower accuracy) add
  "--ac-model-type tri2a" when running the above commands.