Quickly access the latest research papers on large language models from arXiv.
Papers and research related to Large Language Model
QuestBench evaluates large language models' (LLMs) ability to ask clarifying questions in underspecified reasoning tasks, revealing significant challenges in information acquisition despite strong performance in well-defined problems.
The paper presents Unicorn, a novel framework for synthesizing multimodal training data for vision-language models (VLMs) using only text, eliminating the need for costly image-text pairs.
The paper presents an agent-centric personalized clustering framework utilizing multi-modal large language models (MLLMs) to enhance user-specific clustering by effectively traversing relational graphs for better alignment with user interests.
This paper evaluates the use of Multimodal Large Language Models (MLLMs) as assistive technologies for visually impaired users, identifying challenges and areas for improvement in their contextual understanding and usability.
This study investigates the application of large language models (LLMs) for irony detection in 19th-century Latin American newspapers, focusing on enhancing datasets and improving classification tasks.
This research paper presents a method to enhance multilingual capabilities in Visual Language Models (VLMs) by addressing Image-induced Fidelity Loss (IFL) through a continuous multilingual integration strategy.
Niyama is a novel inference serving system designed to optimize resource utilization for Large Language Models (LLMs) by enabling fine-grained Quality-of-Service (QoS) differentiation and efficient workload co-scheduling.
This research paper presents a method to enhance pre-trained text-only large language models (LLMs) with multi-modal generation capabilities while maintaining original performance and efficiency through parameter redundancy exploitation.
This survey evaluates methods for assessing LLM-based agents in multi-turn conversations, establishing a comprehensive framework that categorizes evaluation components and methodologies for effective analysis.
The paper presents Entropy-Guided Sequence Weighting (EGSW), a method that improves exploration-exploitation in Reinforcement Learning for fine-tuning Large Language Models by dynamically weighting outputs based on their advantage and entropy.
CoSIL is a novel LLM-driven method for software issue localization that enhances the accuracy of patch generation by dynamically constructing function call graphs without pre-parsing, achieving significant performance improvements over existing methods.
This research presents a novel approach to typosquatting detection using large language models (LLMs), enhancing cybersecurity by addressing sophisticated domain-based deception tactics.