Quickly access the latest research papers on large language models from arXiv.
Papers and research related to Large Language Model
This study evaluates the capability of Large Language Models (LLMs) for generating domain-specific ontologies, demonstrating their consistent performance across various domains and highlighting their potential for scalable ontology construction.
This research presents a multi-agent framework for evaluating LLM judgments, addressing biases in human evaluation and enhancing the selection of LLM responses through a structured meta-judging process.
This research introduces Ensemble Bayesian Inference (EBI), a method that combines small language models (SLMs) to achieve accuracy levels comparable to large language models (LLMs) in profile matching tasks.
The paper presents INSIGHT, a proof of concept that integrates AI tools, particularly Large Language Models, to enhance student-teacher interactions and personalize education in higher learning environments.
This research presents a Split Conformal Prediction (SCP) framework to mitigate hallucinations in Large Vision-Language Models (LVLMs) for Visual Question Answering (VQA), enhancing reliability in safety-critical applications.
The paper presents a new benchmark for evaluating hallucinations in large language models (LLMs), addressing the challenges of inconsistent definitions and categorizations to enhance user trust in generative AI systems.
This research presents a hybrid model combining Graph Convolutional Networks (GCNs) with Large Language Model (LLM) embeddings to enhance virtual screening in drug discovery, achieving superior predictive performance.
The paper presents HMI, a Hierarchical Knowledge Management system designed to optimize multi-tenant inference in pretrained language models, significantly reducing resource demands while maintaining performance.
This paper presents a novel method that combines large language model (LLM) summaries with topic modeling to identify meaningful topics in Python source code, enhancing software engineering tasks.
The paper 'Replay to Remember' addresses catastrophic forgetting in large language models (LLMs) during continual learning, proposing a lightweight method that combines LoRA and minimal replay for real-time domain adaptation across various fields.
This paper presents a framework for measuring the adversarial robustness of Large Language Models (LLMs), addressing their vulnerability to adversarial inputs and the need for task-specific evaluations.
This position paper discusses the collaborative potential of large language models (LLMs) and smaller models (SMs) to enhance domain-specific tasks, advocating for a synergistic approach to model adaptation and efficiency.
NeuralGrok introduces a gradient-based method to enhance the generalization of transformers in arithmetic tasks, addressing the grokking phenomenon by optimizing gradient transformations for improved model performance.