Sorry, you need to enable JavaScript to visit this website.

Verifying the Long-range Dependency of RNN Language Models

Citation Author(s):
Tzu-Hsuan Tseng, Tzu-Hsuan Yang, Chia-Ping Chen
Submitted by:
Tzu-Hsuan Tseng
Last updated:
21 November 2016 - 10:24am
Document Type:
Presentation Slides
Document Year:
2016
Event:
Presenters:
Tzu-Hsuan, Tseng
Paper Code:
41
 

It has been argued that recurrent neural network language models are better in capturing long-range dependency than n-gram language models. In this paper, we attempt to verify this claim by investigating the prediction accuracy and the perplexity of these language models as a function of word position, i.e., the position of a word in a sentence. It is expected that as word position increases, the advantage of using recurrent neural network language models over n-gram language models will become more and more evident. On the text corpus of Penn Tree Bank (PTB), a recurrent neural network language model outperforms a trigram language model in both perplexity and word prediction. However, on the AMI meeting corpus, a trigram outperforms a recurrent neural network language model.

up
0 users have voted: