Attention in natural language processing

A Galassi, M Lippi, P Torroni - IEEE transactions on neural …, 2020 - ieeexplore.ieee.org
Attention is an increasingly popular mechanism used in a wide range of neural
architectures. The mechanism itself has been realized in a variety of formats. However …

Personalisation within bounds: A risk taxonomy and policy framework for the alignment of large language models with personalised feedback

HR Kirk, B Vidgen, P Röttger, SA Hale - arXiv preprint arXiv:2303.05453, 2023 - arxiv.org
Large language models (LLMs) are used to generate content for a wide range of tasks, and
are set to reach a growing audience in coming years due to integration in product interfaces …

Ftrans: energy-efficient acceleration of transformers using fpga

B Li, S Pandey, H Fang, Y Lyv, J Li, J Chen… - Proceedings of the …, 2020 - dl.acm.org
In natural language processing (NLP), the" Transformer" architecture was proposed as the
first transduction model replying entirely on self-attention mechanisms without using …

Less is more: Attention supervision with counterfactuals for text classification

S Choi, H Park, J Yeo, S Hwang - Proceedings of the 2020 …, 2020 - aclanthology.org
We aim to leverage human and machine intelligence together for attention supervision.
Specifically, we show that human annotation cost can be kept reasonably low, while its …

Deciding whether to ask clarifying questions in large-scale spoken language understanding

JK Kim, G Wang, S Lee, YB Kim - 2021 IEEE Automatic Speech …, 2021 - ieeexplore.ieee.org
A large-scale conversational agent can suffer from understanding user utterances with
various ambiguities such as ASR ambiguity, intent ambiguity, and hypothesis ambiguity …

Pseudo labeling and negative feedback learning for large-scale multi-label domain classification

JK Kim, YB Kim - … 2020-2020 IEEE International Conference on …, 2020 - ieeexplore.ieee.org
In large-scale domain classification, an utterance can be handled by multiple domains with
overlapped capabilities. However, only a limited number of ground-truth domains are …

Meta-supervision for attention using counterfactual estimation

S Choi, H Park, S Hwang - Data Science and Engineering, 2020 - Springer
Neural attention mechanism has been used as a form of explanation for model behavior.
Users can either passively consume explanation or actively disagree with explanation and …

Deep Networks and Knowledge: from Rule Learning to Neural-Symbolic Argument Mining

A Galassi - 2021 - amsdottorato.unibo.it
Deep Learning has revolutionized the whole discipline of machine learning, heavily
impacting fields such as Computer Vision, Natural Language Processing, and other …

[PDF][PDF] Using Domain Knowledge to Improve Machine Translation in Indian Languages

A Gahoi - 2023 - cdn.iiit.ac.in
In this modern world, due to the increased mobility of humans, encountering a foreign
language has become a common challenge for many people around the world. This causes …