A survey of controllable text generation using transformer-based pre-trained language models
Controllable Text Generation (CTG) is an emerging area in the field of natural language
generation (NLG). It is regarded as crucial for the development of advanced text generation …
generation (NLG). It is regarded as crucial for the development of advanced text generation …
A survey of evaluation metrics used for NLG systems
In the last few years, a large number of automatic evaluation metrics have been proposed for
evaluating Natural Language Generation (NLG) systems. The rapid development and …
evaluating Natural Language Generation (NLG) systems. The rapid development and …
Bertscore: Evaluating text generation with bert
We propose BERTScore, an automatic evaluation metric for text generation. Analogously to
common metrics, BERTScore computes a similarity score for each token in the candidate …
common metrics, BERTScore computes a similarity score for each token in the candidate …
MoverScore: Text generation evaluating with contextualized embeddings and earth mover distance
A robust evaluation metric has a profound impact on the development of text generation
systems. A desirable metric compares system output against references based on their …
systems. A desirable metric compares system output against references based on their …
BERT: a review of applications in natural language processing and understanding
MV Koroteev - arXiv preprint arXiv:2103.11943, 2021 - arxiv.org
In this review, we describe the application of one of the most popular deep learning-based
language models-BERT. The paper describes the mechanism of operation of this model, the …
language models-BERT. The paper describes the mechanism of operation of this model, the …
Automatic machine translation evaluation in many languages via zero-shot paraphrasing
B Thompson, M Post - arXiv preprint arXiv:2004.14564, 2020 - arxiv.org
We frame the task of machine translation evaluation as one of scoring machine translation
output with a sequence-to-sequence paraphraser, conditioned on a human reference. We …
output with a sequence-to-sequence paraphraser, conditioned on a human reference. We …
End-to-end transformer-based models in textual-based NLP
A Rahali, MA Akhloufi - AI, 2023 - mdpi.com
Transformer architectures are highly expressive because they use self-attention
mechanisms to encode long-range dependencies in the input sequences. In this paper, we …
mechanisms to encode long-range dependencies in the input sequences. In this paper, we …
Are references really needed? unbabel-IST 2021 submission for the metrics shared task
In this paper, we present the joint contribution of Unbabel and IST to the WMT 2021 Metrics
Shared Task. With this year's focus on Multidimensional Quality Metric (MQM) as the ground …
Shared Task. With this year's focus on Multidimensional Quality Metric (MQM) as the ground …
YiSi-a unified semantic MT quality evaluation and estimation metric for languages with different levels of available resources
C Lo - Proceedings of the Fourth Conference on Machine …, 2019 - aclanthology.org
We present YiSi, a unified automatic semantic machine translation quality evaluation and
estimation metric for languages with different levels of available resources. Underneath the …
estimation metric for languages with different levels of available resources. Underneath the …
[PDF][PDF] Results of the wmt16 metrics shared task
This paper presents the results of the WMT16 Metrics Shared Task. We asked participants of
this task to score the outputs of the MT systems involved in the WMT16 Shared Translation …
this task to score the outputs of the MT systems involved in the WMT16 Shared Translation …