A survey on the memory mechanism of large language model based agents
Large language model (LLM) based agents have recently attracted much attention from the
research and industry communities. Compared with original LLMs, LLM-based agents are …
research and industry communities. Compared with original LLMs, LLM-based agents are …
Dataset Decomposition: Faster LLM Training with Variable Sequence Length Curriculum
Large language models (LLMs) are commonly trained on datasets consisting of fixed-length
token sequences. These datasets are created by randomly concatenating documents of …
token sequences. These datasets are created by randomly concatenating documents of …
LoGra-Med: Long context multi-graph alignment for medical vision-language model
State-of-the-art medical multi-modal large language models (med-MLLM), like LLaVA-Med
or BioMedGPT, leverage instruction-following data in pre-training. However, those models …
or BioMedGPT, leverage instruction-following data in pre-training. However, those models …
Tulip: Token-length upgraded clip
We address the challenge of representing long captions in vision-language models, such as
CLIP. By design these models are limited by fixed, absolute positional encodings, restricting …
CLIP. By design these models are limited by fixed, absolute positional encodings, restricting …
CNNSum: Exploring Long-Conext Summarization with Large Language Models in Chinese Novels
Large Language Models (LLMs) have been well-researched in many long-context tasks.
However, due to high annotation costs, high-quality long-context summary datasets for …
However, due to high annotation costs, high-quality long-context summary datasets for …
Language Models can Self-Lengthen to Generate Long Texts
Recent advancements in Large Language Models (LLMs) have significantly enhanced their
ability to process long contexts, yet a notable gap remains in generating long, aligned …
ability to process long contexts, yet a notable gap remains in generating long, aligned …
Efficiently Exploring Large Language Models for Document-Level Machine Translation with In-context Learning
Large language models (LLMs) exhibit outstanding performance in machine translation via
in-context learning. In contrast to sentence-level translation, document-level translation …
in-context learning. In contrast to sentence-level translation, document-level translation …
A Study on Context Length and Efficient Transformers for Biomedical Image Analysis
SM Hooper, H Xue - arXiv preprint arXiv:2501.00619, 2024 - arxiv.org
Biomedical imaging modalities often produce high-resolution, multi-dimensional images that
pose computational challenges for deep neural networks. These computational challenges …
pose computational challenges for deep neural networks. These computational challenges …
Advancing Bug Detection in Fastjson2 with Large Language Models Driven Unit Test Generation
Z Zhong, S Wang, H Wang, S Wen, H Guan… - arXiv preprint arXiv …, 2024 - arxiv.org
Data-serialization libraries are essential tools in software development, responsible for
converting between programmable data structures and data persistence formats. Among …
converting between programmable data structures and data persistence formats. Among …
The CAP Principle for LLM Serving
We survey the large language model (LLM) serving area to understand the intricate
dynamics between cost-efficiency and accuracy, which is magnified by the growing need for …
dynamics between cost-efficiency and accuracy, which is magnified by the growing need for …