Quickly access the latest research papers on large language models from arXiv.
Papers and research related to Large Language Model
This survey paper examines the role of large language models (LLMs) in automated scholarly paper review (ASPR), highlighting their transformative potential and the challenges they present in academia.
This paper presents a novel architecture that combines Large Language Models (LLMs) with classical automated planners to enhance decision-making in intelligent agents, particularly in formative simulations.
This survey reviews LLM test-time compute via search, addressing task definitions, LLM profiling, and search algorithms to facilitate comparisons among various inference frameworks in the field of artificial intelligence.
The paper presents Mind Evolution, an evolutionary search strategy that enhances inference time compute in Large Language Models, demonstrating superior performance in natural language planning tasks compared to traditional methods.
This research presents a hierarchical autoregressive transformer model that integrates character-level and word-level processing, enhancing robustness and adaptability in natural language processing tasks.
The paper presents a novel Dual Debiasing Algorithm through Model Adaptation (2DAMA) aimed at reducing gender stereotypes in language models while preserving factual gender information for fair language processing.
This research paper presents advancements in NLP for the Uzbek language by evaluating BERT models for part-of-speech tagging and introducing a benchmark dataset, achieving significant accuracy improvements.
The paper introduces TARDIS, a novel method for optimizing large language models (LLMs) by approximating non-linear activation functions with linear ones, achieving significant parameter reduction and improved inference speed.
The paper presents the Attention-Guided Self-Reflection (AGSER) method for zero-shot hallucination detection in Large Language Models (LLMs), significantly improving detection efficacy while reducing computational complexity.
The paper 'Towards Human-Guided, Data-Centric LLM Co-Pilots' presents CliMB-DC, a framework that integrates human expertise with LLMs to address data-centric challenges in machine learning, particularly in healthcare.
This paper surveys advancements in Explainable AI (XAI), focusing on methods that enhance the interpretability of machine learning models, including large language models (LLMs), to foster trust and usability in critical applications.
The paper introduces MultiPruner, a novel pruning method for large pre-trained models that enhances model compression while maintaining accuracy through a multidimensional pruning strategy.