Review
ZHOU Yun1, WANG Ting1, YI Mianzhu2, ZHANG Lupeng3, WANG Zhiyuan1,4
2012, 26(2): 28-35.
All-Words Word Sense Disambiguation (WSD) can be regarded as a sequence labeling problem, and two All-Words WSD methods based on sequence labeling are proposed in this paper, which are based on Hidden Markov Model (HMM) and Maximum Entropy Markov Model (MEMM), respectively. First, we model All-Words WSD using HMM. Since HMM can only exploit lexical observation, we generalize HMM to MEMM by incorporating a large number of non-independent features. For All-Words WSD which is a typical extra-large state problem, the data sparsity and high time complexity seriously hinder the application of HMM and MEMM models. We solve these problems by beam-search Viterbi algorithm and smoothing strategy. Finally, we test our methods on the dataset of All-Words WSD tasks in Senseval-2 and Senseval-3, and achieving a 0.654 F1 value forthe MEMM method which outperforms other methods based on sequence labeling.
Key wordsall-words word sense disambiguation; hidden Markov model; maximum entropy Markov model; very large state problem