HMM Network Instance
(Redirected from Viterbi Lattice)
An HMM network instance is a lattice-based directed conditional probability network that abides by a hidden Markov network metamodel.
- AKA: Hidden Markov Model, Hidden Markov Graph, Viterbi Lattice.
- It can (typically) be composed of:
- It can be associated to Hidden Markov Metamodel.
- It can be a Finite-State Sequence Tagging Model.
- It can be represented with:
- It can be produced by a Hidden Markov Modeling System (that applies an HMM training algorithm to solve an HMM training task)
- See: Undirected Probabilistic Network.
- (Cohen & Hersh, 2005) ⇒ Aaron Michael Cohen, and William R. Hersh. (2005). “A Survey of Current Work in Biomedical Text Mining.” In: Briefings in Bioinformatics 2005 6(1). doi:10.1093/bib/6.1.57
- Zhou et al. trained a hidden Markov model (HMM) on a set of features based on
- Michael Littman's nifty explanatory grid:
|Do we have control
over the state transitons?
|Are the states
MDPMarkov Decision Process
HMMHidden Markov Model
Markov Decision Process
- (Shin, Han et al., 1995) ⇒ Joong-Ho Shin, Young-Soek Han, and Key-Sun Choi. (1995). “A HMM Part-of-Speech Tagger for Korean with wordphrasal Relations". In: Proceedings of Recent Advances in Natural Language Processing (RANLP 1995)
- QUOTE: The trained hidden Markov network reflecting both morpheme and word-phrase relations contains 712 nodes and 28553 edges.
- (Tanaka et al., 1993) ⇒ H. Tanaka & al. (1993). “Classification of Proteins via Successive State Splitting of Hidden Markov Network.” In: ProceedingsW26 in IJCAI93,