nnet-am-decodable-simple.h
17.8 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
// nnet3/nnet-am-decodable-simple.h
// Copyright 2012-2015 Johns Hopkins University (author: Daniel Povey)
// See ../../COPYING for clarification regarding multiple authors
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// THIS CODE IS PROVIDED *AS IS* BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
// KIND, EITHER EXPRESS OR IMPLIED, INCLUDING WITHOUT LIMITATION ANY IMPLIED
// WARRANTIES OR CONDITIONS OF TITLE, FITNESS FOR A PARTICULAR PURPOSE,
// MERCHANTABLITY OR NON-INFRINGEMENT.
// See the Apache 2 License for the specific language governing permissions and
// limitations under the License.
#ifndef KALDI_NNET3_NNET_AM_DECODABLE_SIMPLE_H_
#define KALDI_NNET3_NNET_AM_DECODABLE_SIMPLE_H_
#include <vector>
#include "base/kaldi-common.h"
#include "gmm/am-diag-gmm.h"
#include "hmm/transition-model.h"
#include "itf/decodable-itf.h"
#include "nnet3/nnet-optimize.h"
#include "nnet3/nnet-compute.h"
#include "nnet3/am-nnet-simple.h"
namespace kaldi {
namespace nnet3 {
// See also the decodable object in decodable-simple-looped.h, which is better
// and faster in most situations, including TDNNs and LSTMs (but not for
// BLSTMs).
// Note: the 'simple' in the name means it applies to networks
// for which IsSimpleNnet(nnet) would return true.
struct NnetSimpleComputationOptions {
int32 extra_left_context;
int32 extra_right_context;
int32 extra_left_context_initial;
int32 extra_right_context_final;
int32 frame_subsampling_factor;
int32 frames_per_chunk;
BaseFloat acoustic_scale;
bool debug_computation;
NnetOptimizeOptions optimize_config;
NnetComputeOptions compute_config;
CachingOptimizingCompilerOptions compiler_config;
NnetSimpleComputationOptions():
extra_left_context(0),
extra_right_context(0),
extra_left_context_initial(-1),
extra_right_context_final(-1),
frame_subsampling_factor(1),
frames_per_chunk(50),
acoustic_scale(0.1),
debug_computation(false) {
compiler_config.cache_capacity += frames_per_chunk;
}
void Register(OptionsItf *opts) {
opts->Register("extra-left-context", &extra_left_context,
"Number of frames of additional left-context to add on top "
"of the neural net's inherent left context (may be useful in "
"recurrent setups");
opts->Register("extra-right-context", &extra_right_context,
"Number of frames of additional right-context to add on top "
"of the neural net's inherent right context (may be useful in "
"recurrent setups");
opts->Register("extra-left-context-initial", &extra_left_context_initial,
"If >= 0, overrides the --extra-left-context value at the "
"start of an utterance.");
opts->Register("extra-right-context-final", &extra_right_context_final,
"If >= 0, overrides the --extra-right-context value at the "
"end of an utterance.");
opts->Register("frame-subsampling-factor", &frame_subsampling_factor,
"Required if the frame-rate of the output (e.g. in 'chain' "
"models) is less than the frame-rate of the original "
"alignment.");
opts->Register("acoustic-scale", &acoustic_scale,
"Scaling factor for acoustic log-likelihoods (caution: is a no-op "
"if set in the program nnet3-compute");
opts->Register("frames-per-chunk", &frames_per_chunk,
"Number of frames in each chunk that is separately evaluated "
"by the neural net. Measured before any subsampling, if the "
"--frame-subsampling-factor options is used (i.e. counts "
"input frames");
opts->Register("debug-computation", &debug_computation, "If true, turn on "
"debug for the actual computation (very verbose!)");
// register the optimization options with the prefix "optimization".
ParseOptions optimization_opts("optimization", opts);
optimize_config.Register(&optimization_opts);
// register the compute options with the prefix "computation".
ParseOptions compute_opts("computation", opts);
compute_config.Register(&compute_opts);
}
};
/*
This class handles the neural net computation; it's mostly accessed
via other wrapper classes.
Note: this class used to be called NnetDecodableBase.
It can accept just input features, or input features plus iVectors. */
class DecodableNnetSimple {
public:
/**
This constructor takes features as input, and you can either supply a
single iVector input, estimated in batch-mode ('ivector'), or 'online'
iVectors ('online_ivectors' and 'online_ivector_period', or none at all.
Note: it stores references to all arguments to the constructor, so don't
delete them till this goes out of scope.
@param [in] opts The options class. Warning: it includes an acoustic
weight, whose default is 0.1; you may sometimes want to
change this to 1.0.
@param [in] nnet The neural net that we're going to do the computation with
@param [in] priors Vector of priors-- if supplied and nonempty, we subtract
the log of these priors from the nnet output.
@param [in] feats The input feature matrix.
@param [in] compiler A pointer to the compiler object to use-- this enables the
user to maintain a common object in the calling code that
will cache computations across decodes. Note: the compiler code
has no locking mechanism (and it would be tricky to design one,
as we'd need to lock the individual computations also),
so the calling code has to make sure that if there are
multiple threads, they do not share the same compiler
object.
@param [in] ivector If you are using iVectors estimated in batch mode,
a pointer to the iVector, else NULL.
@param [in] online_ivectors
If you are using iVectors estimated 'online'
a pointer to the iVectors, else NULL.
@param [in] online_ivector_period If you are using iVectors estimated 'online'
(i.e. if online_ivectors != NULL) gives the periodicity
(in frames) with which the iVectors are estimated.
*/
DecodableNnetSimple(const NnetSimpleComputationOptions &opts,
const Nnet &nnet,
const VectorBase<BaseFloat> &priors,
const MatrixBase<BaseFloat> &feats,
CachingOptimizingCompiler *compiler,
const VectorBase<BaseFloat> *ivector = NULL,
const MatrixBase<BaseFloat> *online_ivectors = NULL,
int32 online_ivector_period = 1);
// returns the number of frames of likelihoods. The same as feats_.NumRows()
// in the normal case (but may be less if opts_.frame_subsampling_factor !=
// 1).
inline int32 NumFrames() const { return num_subsampled_frames_; }
inline int32 OutputDim() const { return output_dim_; }
// Gets the output for a particular frame, with 0 <= frame < NumFrames().
// 'output' must be correctly sized (with dimension OutputDim()).
void GetOutputForFrame(int32 frame, VectorBase<BaseFloat> *output);
// Gets the output for a particular frame and pdf_id, with
// 0 <= subsampled_frame < NumFrames(),
// and 0 <= pdf_id < OutputDim().
inline BaseFloat GetOutput(int32 subsampled_frame, int32 pdf_id) {
if (subsampled_frame < current_log_post_subsampled_offset_ ||
subsampled_frame >= current_log_post_subsampled_offset_ +
current_log_post_.NumRows())
EnsureFrameIsComputed(subsampled_frame);
return current_log_post_(subsampled_frame -
current_log_post_subsampled_offset_,
pdf_id);
}
private:
KALDI_DISALLOW_COPY_AND_ASSIGN(DecodableNnetSimple);
// This call is made to ensure that we have the log-probs for this frame
// cached in current_log_post_.
void EnsureFrameIsComputed(int32 subsampled_frame);
// This function does the actual nnet computation; it is called from
// EnsureFrameIsComputed. Any padding at file start/end is done by
// the caller of this function (so the input should exceed the output
// by a suitable amount of context). It puts its output in current_log_post_.
void DoNnetComputation(int32 input_t_start,
const MatrixBase<BaseFloat> &input_feats,
const VectorBase<BaseFloat> &ivector,
int32 output_t_start,
int32 num_subsampled_frames);
// Gets the iVector that will be used for this chunk of frames, if we are
// using iVectors (else does nothing). note: the num_output_frames is
// interpreted as the number of t value, which in the subsampled case is not
// the same as the number of subsampled frames (it would be larger by
// opts_.frame_subsampling_factor).
void GetCurrentIvector(int32 output_t_start,
int32 num_output_frames,
Vector<BaseFloat> *ivector);
// called from constructor
void CheckAndFixConfigs();
// returns dimension of the provided iVectors if supplied, or 0 otherwise.
int32 GetIvectorDim() const;
NnetSimpleComputationOptions opts_;
const Nnet &nnet_;
int32 nnet_left_context_;
int32 nnet_right_context_;
int32 output_dim_;
// the log priors (or the empty vector if the priors are not set in the model)
CuVector<BaseFloat> log_priors_;
const MatrixBase<BaseFloat> &feats_;
// note: num_subsampled_frames_ will equal feats_.NumRows() in the normal case
// when opts_.frame_subsampling_factor == 1.
int32 num_subsampled_frames_;
// ivector_ is the iVector if we're using iVectors that are estimated in batch
// mode.
const VectorBase<BaseFloat> *ivector_;
// online_ivector_feats_ is the iVectors if we're using online-estimated ones.
const MatrixBase<BaseFloat> *online_ivector_feats_;
// online_ivector_period_ helps us interpret online_ivector_feats_; it's the
// number of frames the rows of ivector_feats are separated by.
int32 online_ivector_period_;
// a reference to a compiler passed in via the constructor, which may be
// declared at the top level of the program so that we don't have to recompile
// computations each time.
CachingOptimizingCompiler &compiler_;
// The current log-posteriors that we got from the last time we
// ran the computation.
Matrix<BaseFloat> current_log_post_;
// The time-offset of the current log-posteriors. Note: if
// opts_.frame_subsampling_factor > 1, this will be measured in subsampled
// frames.
int32 current_log_post_subsampled_offset_;
};
class DecodableAmNnetSimple: public DecodableInterface {
public:
/**
This constructor takes features as input, and you can either supply a
single iVector input, estimated in batch-mode ('ivector'), or 'online'
iVectors ('online_ivectors' and 'online_ivector_period', or none at all.
Note: it stores references to all arguments to the constructor, so don't
delete them till this goes out of scope.
@param [in] opts The options class. Warning: it includes an acoustic
weight, whose default is 0.1; you may sometimes want to
change this to 1.0.
@param [in] trans_model The transition model to use. This takes care of the
mapping from transition-id (which is an arg to
LogLikelihood()) to pdf-id (which is used internally).
@param [in] am_nnet The neural net that we're going to do the computation with;
we also get the priors to divide by, if applicable, from here.
@param [in] feats A pointer to the input feature matrix; must be non-NULL.
We
@param [in] ivector If you are using iVectors estimated in batch mode,
a pointer to the iVector, else NULL.
@param [in] ivector If you are using iVectors estimated in batch mode,
a pointer to the iVector, else NULL.
@param [in] online_ivectors
If you are using iVectors estimated 'online'
a pointer to the iVectors, else NULL.
@param [in] online_ivector_period If you are using iVectors estimated 'online'
(i.e. if online_ivectors != NULL) gives the periodicity
(in frames) with which the iVectors are estimated.
@param [in,out] compiler A pointer to a compiler [optional]-- the user
can declare one in the calling code and repeatedly
supply pointers to it, which allows for caching of computations
across consecutive decodes. You'd want to have initialized
the compiler object with as
compiler(am_nnet.GetNnet(), opts.optimize_config).
*/
DecodableAmNnetSimple(const NnetSimpleComputationOptions &opts,
const TransitionModel &trans_model,
const AmNnetSimple &am_nnet,
const MatrixBase<BaseFloat> &feats,
const VectorBase<BaseFloat> *ivector = NULL,
const MatrixBase<BaseFloat> *online_ivectors = NULL,
int32 online_ivector_period = 1,
CachingOptimizingCompiler *compiler = NULL);
virtual BaseFloat LogLikelihood(int32 frame, int32 transition_id);
virtual inline int32 NumFramesReady() const {
return decodable_nnet_.NumFrames();
}
virtual int32 NumIndices() const { return trans_model_.NumTransitionIds(); }
virtual bool IsLastFrame(int32 frame) const {
KALDI_ASSERT(frame < NumFramesReady());
return (frame == NumFramesReady() - 1);
}
private:
KALDI_DISALLOW_COPY_AND_ASSIGN(DecodableAmNnetSimple);
// This compiler object is only used if the 'compiler'
// argument to the constructor is NULL.
CachingOptimizingCompiler compiler_;
DecodableNnetSimple decodable_nnet_;
const TransitionModel &trans_model_;
};
class DecodableAmNnetSimpleParallel: public DecodableInterface {
public:
/**
This decodable object is for use in multi-threaded decoding.
It differs from DecodableAmNnetSimple in two respects:
(1) It doesn't keep around pointers to the features and iVectors;
instead, it creates copies of them (so the caller can
delete the originals).
(2) It doesn't support the user passing in a pointer to the
CachingOptimizingCompiler-- because making that thread safe
would be quite complicated, and in any case multi-threaded
decoding probably makes the most sense when using CPU, and
in that case we don't expect the compilation phase to dominate.
This constructor takes features as input, and you can either supply a
single iVector input, estimated in batch-mode ('ivector'), or 'online'
iVectors ('online_ivectors' and 'online_ivector_period', or none at all.
Note: it stores references to all arguments to the constructor, so don't
delete them till this goes out of scope.
@param [in] opts The options class. Warning: it includes an acoustic
weight, whose default is 0.1; you may sometimes want to
change this to 1.0.
@param [in] trans_model The transition model to use. This takes care of the
mapping from transition-id (which is an arg to
LogLikelihood()) to pdf-id (which is used internally).
@param [in] am_nnet The neural net that we're going to do the computation with;
it may provide priors to divide by.
@param [in] feats A pointer to the input feature matrix; must be non-NULL.
@param [in] ivector If you are using iVectors estimated in batch mode,
a pointer to the iVector, else NULL.
@param [in] online_ivectors
If you are using iVectors estimated 'online'
a pointer to the iVectors, else NULL.
@param [in] online_ivector_period If you are using iVectors estimated 'online'
(i.e. if online_ivectors != NULL) gives the periodicity
(in frames) with which the iVectors are estimated.
*/
DecodableAmNnetSimpleParallel(
const NnetSimpleComputationOptions &opts,
const TransitionModel &trans_model,
const AmNnetSimple &am_nnet,
const MatrixBase<BaseFloat> &feats,
const VectorBase<BaseFloat> *ivector = NULL,
const MatrixBase<BaseFloat> *online_ivectors = NULL,
int32 online_ivector_period = 1);
virtual BaseFloat LogLikelihood(int32 frame, int32 transition_id);
virtual inline int32 NumFramesReady() const {
return decodable_nnet_->NumFrames();
}
virtual int32 NumIndices() const { return trans_model_.NumTransitionIds(); }
virtual bool IsLastFrame(int32 frame) const {
KALDI_ASSERT(frame < NumFramesReady());
return (frame == NumFramesReady() - 1);
}
~DecodableAmNnetSimpleParallel() { DeletePointers(); }
private:
KALDI_DISALLOW_COPY_AND_ASSIGN(DecodableAmNnetSimpleParallel);
void DeletePointers();
CachingOptimizingCompiler compiler_;
const TransitionModel &trans_model_;
Matrix<BaseFloat> *feats_copy_;
Vector<BaseFloat> *ivector_copy_;
Matrix<BaseFloat> *online_ivectors_copy_;
DecodableNnetSimple *decodable_nnet_;
};
} // namespace nnet3
} // namespace kaldi
#endif // KALDI_NNET3_NNET_AM_DECODABLE_SIMPLE_H_