Personal llm agents: Insights and survey about the capability, efficiency and security
Since the advent of personal computing devices, intelligent personal assistants (IPAs) have
been one of the key technologies that researchers and engineers have focused on, aiming …
been one of the key technologies that researchers and engineers have focused on, aiming …
Mathverse: Does your multi-modal llm truly see the diagrams in visual math problems?
The remarkable progress of Multi-modal Large Language Models (MLLMs) has garnered
unparalleled attention, due to their superior performance in visual contexts. However, their …
unparalleled attention, due to their superior performance in visual contexts. However, their …
Mavis: Mathematical visual instruction tuning
Multi-modal Large Language Models (MLLMs) have recently emerged as a significant focus
in academia and industry. Despite their proficiency in general multi-modal scenarios, the …
in academia and industry. Despite their proficiency in general multi-modal scenarios, the …
Referred by multi-modality: A unified temporal transformer for video object segmentation
Recently, video object segmentation (VOS) referred by multi-modal signals, eg, language
and audio, has evoked increasing attention in both industry and academia. It is challenging …
and audio, has evoked increasing attention in both industry and academia. It is challenging …
Integrating large language models into recommendation via mutual augmentation and adaptive aggregation
Conventional recommendation methods have achieved notable advancements by
harnessing collaborative or sequential information from user behavior. Recently, large …
harnessing collaborative or sequential information from user behavior. Recently, large …
Scaling synthetic data creation with 1,000,000,000 personas
We propose a novel persona-driven data synthesis methodology that leverages various
perspectives within a large language model (LLM) to create diverse synthetic data. To fully …
perspectives within a large language model (LLM) to create diverse synthetic data. To fully …
Reft: Reasoning with reinforced fine-tuning
One way to enhance the reasoning capability of Large Language Models (LLMs) is to
conduct Supervised Fine-Tuning (SFT) using Chain-of-Thought (CoT) annotations. This …
conduct Supervised Fine-Tuning (SFT) using Chain-of-Thought (CoT) annotations. This …
Step-dpo: Step-wise preference optimization for long-chain reasoning of llms
Mathematical reasoning presents a significant challenge for Large Language Models
(LLMs) due to the extensive and precise chain of reasoning required for accuracy. Ensuring …
(LLMs) due to the extensive and precise chain of reasoning required for accuracy. Ensuring …
Mathgenie: Generating synthetic data with question back-translation for enhancing mathematical reasoning of llms
Large language models (LLMs) have exhibited great potential in mathematical reasoning.
However, there remains a performance gap in this area between existing open-source …
However, there remains a performance gap in this area between existing open-source …
A survey of neural code intelligence: Paradigms, advances and beyond
Neural Code Intelligence--leveraging deep learning to understand, generate, and optimize
code--holds immense potential for transformative impacts on the whole society. Bridging the …
code--holds immense potential for transformative impacts on the whole society. Bridging the …