Ecomgpt: Instruction-tuning large language models with chain-of-task tasks for e-commerce

Y Li, S Ma, X Wang, S Huang, C Jiang… - Proceedings of the …, 2024 - ojs.aaai.org
Recently, instruction-following Large Language Models (LLMs), represented by ChatGPT,
have exhibited exceptional performance in general Natural Language Processing (NLP) …

Mrrl: Modifying the reference via reinforcement learning for non-autoregressive joint multiple intent detection and slot filling

X Cheng, Z Zhu, B Cao, Q Ye, Y Zou - Findings of the Association …, 2023 - aclanthology.org
With the rise of non-autoregressive approach, some non-autoregressive models for joint
multiple intent detection and slot filling have obtained the promising inference speed …

Seqgpt: An out-of-the-box large language model for open domain sequence understanding

T Yu, C Jiang, C Lou, S Huang, X Wang… - Proceedings of the …, 2024 - ojs.aaai.org
Large language models (LLMs) have shown impressive abilities for open-domain NLP
tasks. However, LLMs are sometimes too footloose for natural language understanding …

MESED: A multi-modal entity set expansion dataset with fine-grained semantic classes and hard negative entities

Y Li, T Lu, HT Zheng, Y Li, S Huang, T Yu… - Proceedings of the …, 2024 - ojs.aaai.org
The Entity Set Expansion (ESE) task aims to expand a handful of seed entities with new
entities belonging to the same semantic class. Conventional ESE methods are based on …

Lateval: An interactive llms evaluation benchmark with incomplete information from lateral thinking puzzles

S Huang, S Ma, Y Li, M Huang, W Zou, W Zhang… - arXiv preprint arXiv …, 2023 - arxiv.org
With the continuous evolution and refinement of LLMs, they are endowed with impressive
logical reasoning or vertical thinking capabilities. But can they think out of the box? Do they …

CLEME: debiasing multi-reference evaluation for grammatical error correction

J Ye, Y Li, Q Zhou, Y Li, S Ma, HT Zheng… - arXiv preprint arXiv …, 2023 - arxiv.org
Evaluating the performance of Grammatical Error Correction (GEC) systems is a challenging
task due to its subjectivity. Designing an evaluation metric that is as objective as possible is …

When llms meet cunning questions: A fallacy understanding benchmark for large language models

Y Li, Q Zhou, Y Luo, S Ma, Y Li, HT Zheng, X Hu… - arXiv preprint arXiv …, 2024 - arxiv.org
Recently, Large Language Models (LLMs) have made remarkable evolutions in language
understanding and generation. Following this, various benchmarks for measuring all kinds …

Towards real-world writing assistance: A chinese character checking benchmark with faked and misspelled characters

Y Li, Z Xu, S Chen, H Huang, Y Li, Y Jiang, Z Li… - arXiv preprint arXiv …, 2023 - arxiv.org
Writing assistance is an application closely related to human life and is also a fundamental
Natural Language Processing (NLP) research field. Its aim is to improve the correctness and …

A Frustratingly Easy Plug-and-Play Detection-and-Reasoning Module for Chinese Spelling Check

H Huang, J Ye, Q Zhou, Y Li, Y Li, F Zhou… - arXiv preprint arXiv …, 2023 - arxiv.org
In recent years, Chinese Spelling Check (CSC) has been greatly improved by designing
task-specific pre-training methods or introducing auxiliary tasks, which mostly solve this task …

Mixedit: Revisiting data augmentation and beyond for grammatical error correction

J Ye, Y Li, Y Li, HT Zheng - arXiv preprint arXiv:2310.11671, 2023 - arxiv.org
Data Augmentation through generating pseudo data has been proven effective in mitigating
the challenge of data scarcity in the field of Grammatical Error Correction (GEC). Various …