Virtual prompt pre-training for prototype-based few-shot relation extraction

K He, Y Huang, R Mao, T Gong, C Li… - Expert Systems with …, 2023 - Elsevier
Prompt tuning with pre-trained language models (PLM) has exhibited outstanding
performance by reducing the gap between pre-training tasks and various downstream …

[PDF][PDF] Virtual prompt pre-training for prototype-based few-shot relation extraction

K He, Y Huang, R Mao, T Gong, C Li… - Expert Systems With …, 2023 - w.sentic.net
Prompt tuning with pre-trained language models (PLM) has exhibited outstanding
performance by reducing the gap between pre-training tasks and various downstream …

[PDF][PDF] Virtual prompt pre-training for prototype-based few-shot relation extraction

K He, Y Huang, R Mao, T Gong, C Li… - Expert Systems With …, 2023 - sentic.net
Prompt tuning with pre-trained language models (PLM) has exhibited outstanding
performance by reducing the gap between pre-training tasks and various downstream …

Virtual prompt pre-training for prototype-based few-shot relation extraction

K He, Y Huang, R Mao, T Gong, C Li… - Expert Systems with …, 2023 - dr.ntu.edu.sg
Prompt tuning with pre-trained language models (PLM) has exhibited outstanding
performance by reducing the gap between pre-training tasks and various downstream …

[PDF][PDF] Virtual prompt pre-training for prototype-based few-shot relation extraction

K He, Y Huang, R Mao, T Gong, C Li… - Expert Systems With …, 2023 - sentic.net
Prompt tuning with pre-trained language models (PLM) has exhibited outstanding
performance by reducing the gap between pre-training tasks and various downstream …

Virtual prompt pre-training for prototype-based few-shot relation extraction

K He, Y Huang, R Mao, T Gong, C Li, E Cambria - 2023 - dl.acm.org
Prompt tuning with pre-trained language models (PLM) has exhibited outstanding
performance by reducing the gap between pre-training tasks and various downstream …

[PDF][PDF] Virtual prompt pre-training for prototype-based few-shot relation extraction

K He, Y Huang, R Mao, T Gong, C Li… - Expert Systems With …, 2023 - ww.sentic.net
Prompt tuning with pre-trained language models (PLM) has exhibited outstanding
performance by reducing the gap between pre-training tasks and various downstream …

[PDF][PDF] Virtual prompt pre-training for prototype-based few-shot relation extraction

K He, Y Huang, R Mao, T Gong, C Li… - Expert Systems With …, 2023 - w.sentic.net
Prompt tuning with pre-trained language models (PLM) has exhibited outstanding
performance by reducing the gap between pre-training tasks and various downstream …

[PDF][PDF] Virtual prompt pre-training for prototype-based few-shot relation extraction

K He, Y Huang, R Mao, T Gong, C Li… - Expert Systems With …, 2023 - ww.sentic.net
Prompt tuning with pre-trained language models (PLM) has exhibited outstanding
performance by reducing the gap between pre-training tasks and various downstream …

[PDF][PDF] Virtual prompt pre-training for prototype-based few-shot relation extraction

K He, Y Huang, R Mao, T Gong, C Li… - Expert Systems With …, 2023 - sentic.net
Prompt tuning with pre-trained language models (PLM) has exhibited outstanding
performance by reducing the gap between pre-training tasks and various downstream …