A primer on neural network models for natural language processing
Y Goldberg - Journal of Artificial Intelligence Research, 2016 - jair.org
Over the past few years, neural networks have re-emerged as powerful machine-learning
models, yielding state-of-the-art results in fields such as image recognition and speech …
models, yielding state-of-the-art results in fields such as image recognition and speech …
[HTML][HTML] Exploring the frontiers of deep learning and natural language processing: A comprehensive overview of key challenges and emerging trends
In the recent past, more than 5 years or so, DL especially the large language models (LLMs)
has generated extensive studies out of a distinctly average downturn field of knowledge …
has generated extensive studies out of a distinctly average downturn field of knowledge …
An empirical evaluation of generic convolutional and recurrent networks for sequence modeling
For most deep learning practitioners, sequence modeling is synonymous with recurrent
networks. Yet recent results indicate that convolutional architectures can outperform …
networks. Yet recent results indicate that convolutional architectures can outperform …
Recent trends in deep learning based natural language processing
Deep learning methods employ multiple processing layers to learn hierarchical
representations of data, and have produced state-of-the-art results in many domains …
representations of data, and have produced state-of-the-art results in many domains …
[图书][B] Neural network methods in natural language processing
Y Goldberg - 2017 - books.google.com
Neural networks are a family of powerful machine learning models and this book focuses on
their application to natural language data. The first half of the book (Parts I and II) covers the …
their application to natural language data. The first half of the book (Parts I and II) covers the …
Natural language processing advancements by deep learning: A survey
Natural Language Processing (NLP) helps empower intelligent machines by enhancing a
better understanding of the human language for linguistic-based human-computer …
better understanding of the human language for linguistic-based human-computer …
Enriching word vectors with subword information
Continuous word representations, trained on large unlabeled corpora are useful for many
natural language processing tasks. Popular models that learn such representations ignore …
natural language processing tasks. Popular models that learn such representations ignore …
[PDF][PDF] End-to-end Sequence Labeling via Bi-directional LSTM-CNNs-CRF
X Ma - arXiv preprint arXiv:1603.01354, 2016 - njuhugn.github.io
State-of-the-art sequence labeling systems traditionally require large amounts of task-
specific knowledge in the form of hand-crafted features and data pre-processing. In this …
specific knowledge in the form of hand-crafted features and data pre-processing. In this …
Character-level convolutional networks for text classification
This article offers an empirical exploration on the use of character-level convolutional
networks (ConvNets) for text classification. We constructed several large-scale datasets to …
networks (ConvNets) for text classification. We constructed several large-scale datasets to …
A unified model for opinion target extraction and target sentiment prediction
Target-based sentiment analysis involves opinion target extraction and target sentiment
classification. However, most of the existing works usually studied one of these two sub …
classification. However, most of the existing works usually studied one of these two sub …