z-logo
Premium
EXPLOITING SYNTACTIC, SEMANTIC, AND LEXICAL REGULARITIES IN LANGUAGE MODELING VIA DIRECTED MARKOV RANDOM FIELDS
Author(s) -
Wang Shaojun,
Wang Shaomin,
Cheng Li,
Greiner Russell,
Schuurmans Dale
Publication year - 2013
Publication title -
computational intelligence
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.353
H-Index - 52
eISSN - 1467-8640
pISSN - 0824-7935
DOI - 10.1111/j.1467-8640.2012.00436.x
Subject(s) - language model , perplexity , computer science , trigram , probabilistic latent semantic analysis , artificial intelligence , natural language processing , smoothing , context (archaeology) , algorithm , paleontology , computer vision , biology
We present a directed Markov random field (MRF) model that combines n ‐gram models, probabilistic context‐free grammars (PCFGs), and probabilistic latent semantic analysis (PLSA) for the purpose of statistical language modeling. Even though the composite directed MRF model potentially has an exponential number of loops and becomes a context‐sensitive grammar, we are nevertheless able to estimate its parameters in cubic time using an efficient modified Expectation‐Maximization (EM) method, the generalized inside–outside algorithm , which extends the inside–outside algorithm to incorporate the effects of the n ‐gram and PLSA language models. We generalize various smoothing techniques to alleviate the sparseness of n ‐gram counts in cases where there are hidden variables. We also derive an analogous algorithm to find the most likely parse of a sentence and to calculate the probability of initial subsequence of a sentence, all generated by the composite language model. Our experimental results on the Wall Street Journal corpus show that we obtain significant reductions in perplexity compared to the state‐of‐the‐art baseline trigram model with Good–Turing and Kneser–Ney smoothing techniques.

This content is not available in your region!

Continue researching here.

Having issues? You can contact us here