Sorry, you need to enable JavaScript to visit this website.

A RANDOM GOSSIP BMUF PROCESS FOR NEURAL LANGUAGE MODELING

Citation Author(s):
Yiheng Huang; Jinchuan Tian; Lei Han; Guangsen Wang; Xingchen Song; Dan Su; Dong Yu
Submitted by:
Yiheng Huang
Last updated:
12 March 2020 - 9:03am
Document Type:
Presentation Slides
Document Year:
2020
Event:
Presenters:
Yiheng Huang, Jinchuan Tian
Paper Code:
1097
 

Neural network language model (NNLM) is an essential component of industrial ASR systems. One important challenge of training an NNLM is to leverage between scaling the learning process and handling big data. Conventional approaches such as block momentum provides a blockwise model update filtering (BMUF) process and achieves almost linear speedups with no performance degradation for speech recognition. However, it needs to calculate the model average from all computing nodes (e.g., GPUs) and when the number of computing nodes is large, the learning suffers from the severe communication latency. As a consequence, BMUF is not suitable under restricted network conditions. In this paper, we present a decentralized BMUF process, in which the model is split into different components, each of which is updated by communicating to some randomly chosen neighbor nodes with the same component, followed by a BMUF-like
process. We apply this method to several LSTM language modeling tasks. Experimental results show that our approach achieves consistently better performance than conventional BMUF. In particular, we obtain a lower perplexity than the single-GPU baseline on the wiki-text-103 benchmark using 4 GPUs. In addition, no performance degradation is observed when scaling to 8 and 16 GPUs.

up
0 users have voted: