Attention in natural language processing
Attention is an increasingly popular mechanism used in a wide range of neural
architectures. The mechanism itself has been realized in a variety of formats. However …
architectures. The mechanism itself has been realized in a variety of formats. However …
Personalisation within bounds: A risk taxonomy and policy framework for the alignment of large language models with personalised feedback
Large language models (LLMs) are used to generate content for a wide range of tasks, and
are set to reach a growing audience in coming years due to integration in product interfaces …
are set to reach a growing audience in coming years due to integration in product interfaces …
Ftrans: energy-efficient acceleration of transformers using fpga
In natural language processing (NLP), the" Transformer" architecture was proposed as the
first transduction model replying entirely on self-attention mechanisms without using …
first transduction model replying entirely on self-attention mechanisms without using …
Less is more: Attention supervision with counterfactuals for text classification
We aim to leverage human and machine intelligence together for attention supervision.
Specifically, we show that human annotation cost can be kept reasonably low, while its …
Specifically, we show that human annotation cost can be kept reasonably low, while its …
Deciding whether to ask clarifying questions in large-scale spoken language understanding
A large-scale conversational agent can suffer from understanding user utterances with
various ambiguities such as ASR ambiguity, intent ambiguity, and hypothesis ambiguity …
various ambiguities such as ASR ambiguity, intent ambiguity, and hypothesis ambiguity …
Pseudo labeling and negative feedback learning for large-scale multi-label domain classification
JK Kim, YB Kim - … 2020-2020 IEEE International Conference on …, 2020 - ieeexplore.ieee.org
In large-scale domain classification, an utterance can be handled by multiple domains with
overlapped capabilities. However, only a limited number of ground-truth domains are …
overlapped capabilities. However, only a limited number of ground-truth domains are …
Meta-supervision for attention using counterfactual estimation
Neural attention mechanism has been used as a form of explanation for model behavior.
Users can either passively consume explanation or actively disagree with explanation and …
Users can either passively consume explanation or actively disagree with explanation and …
Deep Networks and Knowledge: from Rule Learning to Neural-Symbolic Argument Mining
A Galassi - 2021 - amsdottorato.unibo.it
Deep Learning has revolutionized the whole discipline of machine learning, heavily
impacting fields such as Computer Vision, Natural Language Processing, and other …
impacting fields such as Computer Vision, Natural Language Processing, and other …
[PDF][PDF] Using Domain Knowledge to Improve Machine Translation in Indian Languages
A Gahoi - 2023 - cdn.iiit.ac.in
In this modern world, due to the increased mobility of humans, encountering a foreign
language has become a common challenge for many people around the world. This causes …
language has become a common challenge for many people around the world. This causes …