论文标题

在神经语言建模中调查大幅度毛细管软性

Investigation of Large-Margin Softmax in Neural Language Modeling

论文作者

Huo, Jingjing, Gao, Yingbo, Wang, Weiyue, Schlüter, Ralf, Ney, Hermann

论文摘要

为了鼓励可训练的特征向量之间的阶层内紧凑性和类间的可分离性,开发了大型固定软性方法,并广泛应用于面部识别社区。据报道,将大规模利润的概念引入了SoftMax,具有良好的特性,例如增强的判别能力,过度拟合和定义明确的几何直觉。如今,通常使用SoftMax和横熵的神经网络接近语言建模。在这项工作中,我们很想知道在神经语言模型中引入大规模细纹是否会改善困惑性,因此在自动语音识别中会提高单词错误率。具体而言,我们首先在面部识别的先前工作后首先实施和测试各种常规边缘。为了解决自然语言数据的分布,我们然后比较了单词矢量规范尺度的不同策略。之后,我们将最佳的规范设置与各种边缘结合使用,并在自动语音识别中进行了神经语言模型。我们发现,尽管困惑性略有恶化,但具有较大毛细笔柔性的神经语言模型可以产生与标准软马克斯基线相似的单词错误率。最后,通过可视化单词向量来分析预期的边缘,这表明句法和语义关系也得到了保留。

To encourage intra-class compactness and inter-class separability among trainable feature vectors, large-margin softmax methods are developed and widely applied in the face recognition community. The introduction of the large-margin concept into the softmax is reported to have good properties such as enhanced discriminative power, less overfitting and well-defined geometric intuitions. Nowadays, language modeling is commonly approached with neural networks using softmax and cross entropy. In this work, we are curious to see if introducing large-margins to neural language models would improve the perplexity and consequently word error rate in automatic speech recognition. Specifically, we first implement and test various types of conventional margins following the previous works in face recognition. To address the distribution of natural language data, we then compare different strategies for word vector norm-scaling. After that, we apply the best norm-scaling setup in combination with various margins and conduct neural language models rescoring experiments in automatic speech recognition. We find that although perplexity is slightly deteriorated, neural language models with large-margin softmax can yield word error rate similar to that of the standard softmax baseline. Finally, expected margins are analyzed through visualization of word vectors, showing that the syntactic and semantic relationships are also preserved.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源