A primer on neural network models for natural language processing

Y Goldberg - Journal of Artificial Intelligence Research, 2016 - jair.org
Over the past few years, neural networks have re-emerged as powerful machine-learning
models, yielding state-of-the-art results in fields such as image recognition and speech …

[HTML][HTML] Exploring the frontiers of deep learning and natural language processing: A comprehensive overview of key challenges and emerging trends

W Khan, A Daud, K Khan, S Muhammad… - Natural Language …, 2023 - Elsevier
In the recent past, more than 5 years or so, DL especially the large language models (LLMs)
has generated extensive studies out of a distinctly average downturn field of knowledge …

An empirical evaluation of generic convolutional and recurrent networks for sequence modeling

S Bai, JZ Kolter, V Koltun - arXiv preprint arXiv:1803.01271, 2018 - arxiv.org
For most deep learning practitioners, sequence modeling is synonymous with recurrent
networks. Yet recent results indicate that convolutional architectures can outperform …

Recent trends in deep learning based natural language processing

T Young, D Hazarika, S Poria… - ieee Computational …, 2018 - ieeexplore.ieee.org
Deep learning methods employ multiple processing layers to learn hierarchical
representations of data, and have produced state-of-the-art results in many domains …

[图书][B] Neural network methods in natural language processing

Y Goldberg - 2017 - books.google.com
Neural networks are a family of powerful machine learning models and this book focuses on
their application to natural language data. The first half of the book (Parts I and II) covers the …

Natural language processing advancements by deep learning: A survey

A Torfi, RA Shirvani, Y Keneshloo, N Tavaf… - arXiv preprint arXiv …, 2020 - arxiv.org
Natural Language Processing (NLP) helps empower intelligent machines by enhancing a
better understanding of the human language for linguistic-based human-computer …

Enriching word vectors with subword information

P Bojanowski, E Grave, A Joulin… - Transactions of the …, 2017 - direct.mit.edu
Continuous word representations, trained on large unlabeled corpora are useful for many
natural language processing tasks. Popular models that learn such representations ignore …

[PDF][PDF] End-to-end Sequence Labeling via Bi-directional LSTM-CNNs-CRF

X Ma - arXiv preprint arXiv:1603.01354, 2016 - njuhugn.github.io
State-of-the-art sequence labeling systems traditionally require large amounts of task-
specific knowledge in the form of hand-crafted features and data pre-processing. In this …

Character-level convolutional networks for text classification

X Zhang, J Zhao, Y LeCun - Advances in neural information …, 2015 - proceedings.neurips.cc
This article offers an empirical exploration on the use of character-level convolutional
networks (ConvNets) for text classification. We constructed several large-scale datasets to …

A unified model for opinion target extraction and target sentiment prediction

X Li, L Bing, P Li, W Lam - Proceedings of the AAAI conference on …, 2019 - ojs.aaai.org
Target-based sentiment analysis involves opinion target extraction and target sentiment
classification. However, most of the existing works usually studied one of these two sub …