A survey on neural data-to-text generation
Data-to-text Generation (D2T) aims to generate textual natural language statements that can
fluently and precisely describe the structured data such as graphs, tables, and meaning …
fluently and precisely describe the structured data such as graphs, tables, and meaning …
Conditional generation with a question-answering blueprint
The ability to convey relevant and faithful information is critical for many tasks in conditional
generation and yet remains elusive for neural seq-to-seq models whose outputs often reveal …
generation and yet remains elusive for neural seq-to-seq models whose outputs often reveal …
Bridging the structural gap between encoding and decoding for data-to-text generation
Generating sequential natural language descriptions from graph-structured data (eg,
knowledge graph) is challenging, partly because of the structural differences between the …
knowledge graph) is challenging, partly because of the structural differences between the …
Have your text and use it too! end-to-end neural data-to-text generation with semantic fidelity
End-to-end neural data-to-text (D2T) generation has recently emerged as an alternative to
pipeline-based architectures. However, it has faced challenges in generalizing to new …
pipeline-based architectures. However, it has faced challenges in generalizing to new …
Neural pipeline for zero-shot data-to-text generation
In data-to-text (D2T) generation, training on in-domain data leads to overfitting to the data
representation and repeating training data noise. We examine how to avoid finetuning …
representation and repeating training data noise. We examine how to avoid finetuning …
Data-to-text generation with variational sequential planning
We consider the task of data-to-text generation, which aims to create textual output from non-
linguistic input. We focus on generating long-form text, that is, documents with multiple …
linguistic input. We focus on generating long-form text, that is, documents with multiple …
Control prefixes for parameter-efficient text generation
Prefix-tuning is a powerful lightweight technique for adapting a large pre-trained language
model to a downstream application. However, it uses the same dataset-level tuned prompt …
model to a downstream application. However, it uses the same dataset-level tuned prompt …
Towards faithfulness in open domain table-to-text generation from an entity-centric view
In open domain table-to-text generation, we notice the unfaithful generation usually contains
hallucinated entities which can not be aligned to any input table record. We thus try to …
hallucinated entities which can not be aligned to any input table record. We thus try to …
Data-to-text generation with iterative text editing
We present a novel approach to data-to-text generation based on iterative text editing. Our
approach maximizes the completeness and semantic accuracy of the output text while …
approach maximizes the completeness and semantic accuracy of the output text while …
Controllable meaning representation to text generation: Linearization and data augmentation strategies
We study the degree to which neural sequence-to-sequence models exhibit fine-grained
controllability when performing natural language generation from a meaning representation …
controllability when performing natural language generation from a meaning representation …