Position models and language modeling

Abstract : In statistical language modelling the classic model used is $n$-gram. This model is not able however to capture long term dependencies, \emph{i.e.} dependencies larger than $n$. An alternative to this model is the probabilistic automaton. Unfortunately, it appears that preliminary experiments on the use of this model in language modelling is not yet competitive, partly because it tries to model too long term dependencies. We propose here to improve the use of this model by restricting the dependency to a more reasonable value. Experiments shows an improvement of 45\% reduction in the perplexity obtained on the Wall Street Journal language modeling task.
Document type :
Conference papers
Complete list of metadatas

https://hal-ujm.archives-ouvertes.fr/ujm-00322820
Contributor : Franck Thollard <>
Submitted on : Monday, March 9, 2009 - 12:01:04 PM
Last modification on : Wednesday, July 25, 2018 - 2:05:30 PM
Long-term archiving on : Friday, June 4, 2010 - 11:35:07 AM

File

zdiziobeck_thollard_sspr.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : ujm-00322820, version 1

Collections

Citation

Arnaud Zdziobeck, Franck Thollard. Position models and language modeling. Structural and Syntactic Pattern Recognition and Statistical Techniques in Pattern Recognition, Dec 2008, Orlando, United States. pp.76-85. ⟨ujm-00322820⟩

Share

Metrics

Record views

115

Files downloads

137