Home | Trees | Indices | Help |
|
---|
|
object --+ | HiddenMarkovModelTrainer
Algorithms for learning HMM parameters from training data. These include both supervised learning (MLE) and unsupervised learning (Baum-Welch).
|
|||
|
|||
HiddenMarkovModelTagger |
|
||
HiddenMarkovModelTagger |
|
||
HiddenMarkovModelTagger |
|
||
Inherited from |
|
|||
Inherited from |
|
Creates an HMM trainer to induce an HMM with the given states and output symbol alphabet. A supervised and unsupervised training method may be used. If either of the states or symbols are not given, these may be derived from supervised training.
|
Trains the HMM using both (or either of) supervised and unsupervised techniques.
|
Trains the HMM using the Baum-Welch algorithm to maximise the probability of the data sequence. This is a variant of the EM algorithm, and is unsupervised in that it doesn't need the state sequences for the symbols. The code is based on 'A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition', Lawrence Rabiner, IEEE, 1989.
|
Supervised training maximising the joint probability of the symbol and state sequences. This is done via collecting frequencies of transitions between states, symbol observations while within each state and which states start a sentence. These frequency distributions are then normalised into probability estimates, which can be smoothed if desired.
|
Home | Trees | Indices | Help |
|
---|
Generated by Epydoc 3.0beta1 on Wed Aug 27 15:08:57 2008 | http://epydoc.sourceforge.net |