Group sparse CNNs for question classification with answer sets
Mingbo Ma, Liang Huang, et al.
ACL 2017
This paper is an empirical study of the distributed deep learning for question answering subtasks: answer selection and question classification. Comparison studies of SGD, MSGD, ADADELTA, ADAGRAD, ADAM/ADAMAX, RM-SPROP, DOWNPOUR and EASGD/EAMSGD algorithms have been presented. Experimental results show that the distributed framework based on the message passing interface can accelerate the convergence speed at a sublinear scale. This paper demonstrates the importance of distributed training. For example, with 48 workers, a 24 × speedup is achievable for the answer selection task and running time is decreased from 138.2 hours to 5.81 hours, which will increase the productivity significantly.
Mingbo Ma, Liang Huang, et al.
ACL 2017
Bowen Zhou, Bing Xiang, et al.
SSST 2008
Mo Yu, Wenpeng Yin, et al.
ACL 2017
Jia Cui, Yonggang Deng, et al.
ASRU 2009