Deep learning--based text classification: a comprehensive review

S Minaee, N Kalchbrenner, E Cambria… - ACM computing …, 2021 - dl.acm.org
Deep learning--based models have surpassed classical machine learning--based
approaches in various text classification tasks, including sentiment analysis, news …

Information retrieval: recent advances and beyond

KA Hambarde, H Proenca - IEEE Access, 2023 - ieeexplore.ieee.org
This paper provides an extensive and thorough overview of the models and techniques
utilized in the first and second stages of the typical information retrieval processing chain …

Multi-task deep neural networks for natural language understanding

X Liu, P He, W Chen, J Gao - arXiv preprint arXiv:1901.11504, 2019 - arxiv.org
In this paper, we present a Multi-Task Deep Neural Network (MT-DNN) for learning
representations across multiple natural language understanding (NLU) tasks. MT-DNN not …

Attention, please! A survey of neural attention models in deep learning

A de Santana Correia, EL Colombini - Artificial Intelligence Review, 2022 - Springer
In humans, Attention is a core property of all perceptual and cognitive operations. Given our
limited ability to process competing sources, attention mechanisms select, modulate, and …

Semantics-aware BERT for language understanding

Z Zhang, Y Wu, H Zhao, Z Li, S Zhang, X Zhou… - Proceedings of the …, 2020 - ojs.aaai.org
The latest work on language representations carefully integrates contextualized features into
language model training, which enables a series of success especially in various machine …

Knowledge-grounded dialogue generation with pre-trained language models

X Zhao, W Wu, C Xu, C Tao, D Zhao, R Yan - arXiv preprint arXiv …, 2020 - arxiv.org
We study knowledge-grounded dialogue generation with pre-trained language models. To
leverage the redundant external knowledge under capacity constraint, we propose …

The natural language decathlon: Multitask learning as question answering

B McCann, NS Keskar, C Xiong, R Socher - arXiv preprint arXiv …, 2018 - arxiv.org
Deep learning has improved performance on many natural language processing (NLP)
tasks individually. However, general NLP models cannot emerge within a paradigm that …

Exploiting edge features for graph neural networks

L Gong, Q Cheng - … of the IEEE/CVF conference on …, 2019 - openaccess.thecvf.com
Edge features contain important information about graphs. However, current state-of-the-art
neural network models designed for graph learning, eg, graph convolutional networks …

Simple and effective text matching with richer alignment features

R Yang, J Zhang, X Gao, F Ji, H Chen - arXiv preprint arXiv:1908.00300, 2019 - arxiv.org
In this paper, we present a fast and strong neural approach for general purpose text
matching applications. We explore what is sufficient to build a fast and well-performed text …

Deep reinforcement learning for sequence-to-sequence models

Y Keneshloo, T Shi, N Ramakrishnan… - IEEE transactions on …, 2019 - ieeexplore.ieee.org
In recent times, sequence-to-sequence (seq2seq) models have gained a lot of popularity
and provide state-of-the-art performance in a wide variety of tasks, such as machine …