Back to Search Start Over

基于神经网络语言模型的动态层序Softmax训练算法.

Authors :
杨鹤标
胡惊涛
刘芳
Source :
Journal of Jiangsu University (Natural Science Edition) / Jiangsu Daxue Xuebao (Ziran Kexue Ban). 2020, Vol. 41 Issue 1, p67-80. 7p.
Publication Year :
2020

Abstract

To solve the problems of hierarchical Softmax algorithm in the training process of word vectors with inability of incremental training and inefficient training of massive data, the dynamic hierarchical Softmax algorithm was proposed. By the incremental loading of data samples, an adaptive Huffman coding tree was dynamically constructed by the node adjustment replacement method. To avoid the oscillatory decline of loss function due to the small sample size, the first-order and the second-order moment estimations of the gradient were used to dynamically adjust the parameters update direction and learning rate. The weight variation range and the convergence training network error were reduced by the gradient descent algorithm to improve the training efficiency of the word vector from massive data. The Wikipedia Chinese corpus was adopted as the data to test the training efficiency and quality. The experimental results show that the dynamic hierarchical Softmax algorithm can significantly improve the training efficiency and ensure the quality of word vector training. When the incremental samples are from 10 kB to 1 MB, the training speed is increased about 30 times, which can effectively shorten the training period. [ABSTRACT FROM AUTHOR]

Details

Language :
Chinese
ISSN :
16717775
Volume :
41
Issue :
1
Database :
Academic Search Index
Journal :
Journal of Jiangsu University (Natural Science Edition) / Jiangsu Daxue Xuebao (Ziran Kexue Ban)
Publication Type :
Academic Journal
Accession number :
141165922
Full Text :
https://doi.org/10.3969/j.issn.1671-7775.2020.01.011