Strategies for training large vocabulary neural language models

W Chen, D Grangier, M Auli - arXiv preprint arXiv:1512.04906, 2015 - arxiv.org
Training neural network language models over large vocabularies is still computationally
very costly compared to count-based models such as Kneser-Ney. At the same time, neural …

A deep language model for software code

HK Dam, T Tran, T Pham - arXiv preprint arXiv:1608.02715, 2016 - arxiv.org
Existing language models such as n-grams for software code often fail to capture a long
context where dependent code elements scatter far apart. In this paper, we propose a novel …

Top-down tree long short-term memory networks

X Zhang, L Lu, M Lapata - arXiv preprint arXiv:1511.00060, 2015 - arxiv.org
Long Short-Term Memory (LSTM) networks, a type of recurrent neural network with a more
complex computational unit, have been successfully applied to a variety of sequence …

[PDF][PDF] Recurrent neural network language model adaptation for multi-genre broadcast speech recognition

X Chen, T Tan, X Liu, P Lanchantin, M Wan… - … Annual Conference of …, 2015 - academia.edu
Recurrent neural network language models (RNNLMs) have recently become increasingly
popular for many applications including speech recognition. In previous research RNNLMs …

CUED-RNNLM—An open-source toolkit for efficient training and evaluation of recurrent neural network language models

X Chen, X Liu, Y Qian, MJF Gales… - … on acoustics, speech …, 2016 - ieeexplore.ieee.org
In recent years, recurrent neural network language models (RNNLMs) have become
increasingly popular for a range of applications including speech recognition. However, the …

End-to-end speaker verification via curriculum bipartite ranking weighted binary cross-entropy

Z Bai, J Wang, XL Zhang, J Chen - IEEE/ACM Transactions on …, 2022 - ieeexplore.ieee.org
End-to-end speaker verification achieves the verification through estimating directly the
similarity score between a pair of utterances, which is formulated as a binary (ie, target …

Acoustic and textual data augmentation for improved ASR of code-switching speech

E Yılmaz, H Heuvel, DA van Leeuwen - arXiv preprint arXiv:1807.10945, 2018 - arxiv.org
In this paper, we describe several techniques for improving the acoustic and language
model of an automatic speech recognition (ASR) system operating on code-switching (CS) …

Deep learning based an efficient hybrid prediction model for Covid-19 cross-country spread among E7 and G7 countries

A Utku - Decision Making: Applications in Management and …, 2023 - dmame-journal.org
The COVID-19 pandemic has caused the death of many people around the world and has
also caused economic problems for all countries in the world. In the literature, there are …

Blackout: Speeding up recurrent neural network language models with very large vocabularies

S Ji, SVN Vishwanathan, N Satish… - arXiv preprint arXiv …, 2015 - arxiv.org
We propose BlackOut, an approximation algorithm to efficiently train massive recurrent
neural network language models (RNNLMs) with million word vocabularies. BlackOut is …

[PDF][PDF] Convolutional neural network language models

NQ Pham, G Kruszewski, G Boleda - Proceedings of the 2016 …, 2016 - aclanthology.org
Abstract Convolutional Neural Networks (CNNs) have shown to yield very strong results in
several Computer Vision tasks. Their application to language has received much less …