make_unigram_grammar.pl
1.77 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
#!/usr/bin/env perl
# Copyright 2010-2012 Microsoft Corporation Johns Hopkins University (Author: Daniel Povey)
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# THIS CODE IS PROVIDED *AS IS* BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, EITHER EXPRESS OR IMPLIED, INCLUDING WITHOUT LIMITATION ANY IMPLIED
# WARRANTIES OR CONDITIONS OF TITLE, FITNESS FOR A PARTICULAR PURPOSE,
# MERCHANTABLITY OR NON-INFRINGEMENT.
# See the Apache 2 License for the specific language governing permissions and
# limitations under the License.
# This script is used in discriminative training.
# This script makes a simple unigram-loop version of G.fst
# using a unigram grammar estimated from some training transcripts.
# This is for MMI training.
# We don't have any silences in G.fst; these are supplied by the
# optional silences in the lexicon.
# Note: the symbols in the transcripts become the input and output
# symbols of G.txt; these can be numeric or not.
if(@ARGV != 0) {
die "Usage: make_unigram_grammar.pl < text-transcripts > G.txt"
}
$totcount = 0;
$nl = 0;
while (<>) {
@A = split(" ", $_);
foreach $a (@A) {
$count{$a}++;
$totcount++;
}
$nl++;
$totcount++; # Treat end-of-sentence as a symbol for purposes of
# $totcount, so the grammar is properly stochastic. This doesn't
# become </s>, it just becomes the final-prob.
}
foreach $a (keys %count) {
$prob = $count{$a} / $totcount;
$cost = -log($prob); # Negated natural-log probs.
print "0\t0\t$a\t$a\t$cost\n";
}
# Zero final-cost.
$final_prob = $nl / $totcount;
$final_cost = -log($final_prob);
print "0\t$final_cost\n";