"large language models" Papers

986 papers found • Page 6 of 20

Fiddler: CPU-GPU Orchestration for Fast Inference of Mixture-of-Experts Models

Keisuke Kamahori, Tian Tang, Yile Gu et al.

ICLR 2025arXiv:2402.07033
48
citations

Filling Memory Gaps: Enhancing Continual Semantic Parsing via SQL Syntax Variance-Guided LLMs Without Real Data Replay

Ruiheng Liu, Jinyu Zhang, Yanqi Song et al.

AAAI 2025paperarXiv:2412.07246
4
citations

Finding and Reactivating Post-Trained LLMs' Hidden Safety Mechanisms

Mingjie Li, Wai Man Si, Michael Backes et al.

NEURIPS 2025
1
citations

Fine-tuning can Help Detect Pretraining Data from Large Language Models

Hengxiang Zhang, Songxin Zhang, Bingyi Jing et al.

ICLR 2025arXiv:2410.10880
4
citations

Fleet of Agents: Coordinated Problem Solving with Large Language Models

Lars Klein, Nearchos Potamitis, Roland Aydin et al.

ICML 2025arXiv:2405.06691
3
citations

FlexPrefill: A Context-Aware Sparse Attention Mechanism for Efficient Long-Sequence Inference

Xunhao Lai, Jianqiao Lu, Yao Luo et al.

ICLR 2025arXiv:2502.20766
62
citations

Flick: Empowering Federated Learning with Commonsense Knowledge

Ran Zhu, Mingkun Yang, Shiqiang Wang et al.

NEURIPS 2025

FlowerTune: A Cross-Domain Benchmark for Federated Fine-Tuning of Large Language Models

Yan Gao, Massimo R. Scamarcia, Javier Fernandez-Marques et al.

NEURIPS 2025arXiv:2506.02961
4
citations

FlowPrune: Accelerating Attention Flow Calculation by Pruning Flow Network

Shuo Xu, Yu Chen, Shuxia Lin et al.

NEURIPS 2025

FoGE: Fock Space inspired encoding for graph prompting

Takis Chytas, Rudrasis Chakraborty, Vikas Singh

NEURIPS 2025arXiv:2507.02937

ForecastBench: A Dynamic Benchmark of AI Forecasting Capabilities

Ezra Karger, Houtan Bastani, Chen Yueh-Han et al.

ICLR 2025arXiv:2409.19839
34
citations

Forking Paths in Neural Text Generation

Eric Bigelow, Ari Holtzman, Hidenori Tanaka et al.

ICLR 2025arXiv:2412.07961
20
citations

FP4 All the Way: Fully Quantized Training of Large Language Models

Brian Chmiel, Maxim Fishman, Ron Banner et al.

NEURIPS 2025spotlight

From Artificial Needles to Real Haystacks: Improving Retrieval Capabilities in LLMs by Finetuning on Synthetic Data

Zheyang Xiong, Vasilis Papageorgiou, Kangwook Lee et al.

ICLR 2025arXiv:2406.19292
19
citations

From Attention to Activation: Unraveling the Enigmas of Large Language Models

Prannay Kaul, Chengcheng Ma, Ismail Elezi et al.

ICLR 2025arXiv:2410.17174
8
citations

From Euler to AI: Unifying Formulas for Mathematical Constants

Tomer Raz, Michael Shalyt, Elyasheev Leibtag et al.

NEURIPS 2025arXiv:2502.17533
1
citations

From Low Rank Gradient Subspace Stabilization to Low-Rank Weights: Observations, Theories, and Applications

Ajay Jaiswal, Yifan Wang, Lu Yin et al.

ICML 2025arXiv:2407.11239
20
citations

From Programs to Poses: Factored Real-World Scene Generation via Learned Program Libraries

Joy Hsu, Emily Jin, Jiajun Wu et al.

NEURIPS 2025arXiv:2510.10292
1
citations

Functional Homotopy: Smoothing Discrete Optimization via Continuous Parameters for LLM Jailbreak Attacks

Zi Wang, Divyam Anshumaan, Ashish Hooda et al.

ICLR 2025arXiv:2410.04234
4
citations

General-Reasoner: Advancing LLM Reasoning Across All Domains

Xueguang Ma, Qian Liu, Dongfu Jiang et al.

NEURIPS 2025arXiv:2505.14652
86
citations

General Scene Adaptation for Vision-and-Language Navigation

Haodong Hong, Yanyuan Qiao, Sen Wang et al.

ICLR 2025arXiv:2501.17403
10
citations

Generating Computational Cognitive models using Large Language Models

Milena Rmus, Akshay Kumar Jagadish, Marvin Mathony et al.

NEURIPS 2025oralarXiv:2502.00879
5
citations

Generating Traffic Scenarios via In-Context Learning to Learn Better Motion Planner

Aizierjiang Aiersilan

AAAI 2025paperarXiv:2412.18086
4
citations

Generative Monoculture in Large Language Models

Fan Wu, Emily Black, Varun Chandrasekaran

ICLR 2025arXiv:2407.02209
12
citations

Generator-Mediated Bandits: Thompson Sampling for GenAI-Powered Adaptive Interventions

Marc Brooks, Gabriel Durham, Kihyuk Hong et al.

NEURIPS 2025arXiv:2505.16311
1
citations

GeoCAD: Local Geometry-Controllable CAD Generation with Large Language Models

Zhanwei Zhang, kaiyuan liu, Junjie Liu et al.

NEURIPS 2025arXiv:2506.10337
2
citations

GlyphDraw2: Automatic Generation of Complex Glyph Posters with Diffusion Models and Large Language Models

Jian Ma, Yonglin Deng, Chen Chen et al.

AAAI 2025paperarXiv:2407.02252
23
citations

GnnXemplar: Exemplars to Explanations - Natural Language Rules for Global GNN Interpretability

Burouj Armgaan, Eshan Jain, Harsh Pandey et al.

NEURIPS 2025oralarXiv:2509.18376
2
citations

GOFA: A Generative One-For-All Model for Joint Graph Language Modeling

Lecheng Kong, Jiarui Feng, Hao Liu et al.

ICLR 2025arXiv:2407.09709
30
citations

Gradient Multi-Normalization for Efficient LLM Training

Meyer Scetbon, Chao Ma, Wenbo Gong et al.

NEURIPS 2025
3
citations

Grammars of Formal Uncertainty: When to Trust LLMs in Automated Reasoning Tasks

Debargha Ganguly, Vikash Singh, Sreehari Sankar et al.

NEURIPS 2025arXiv:2505.20047
5
citations

GraphChain: Large Language Models for Large-scale Graph Analysis via Tool Chaining

Chunyu Wei, Wenji Hu, Xingjia Hao et al.

NEURIPS 2025arXiv:2511.00457
1
citations

GRIFFIN: Effective Token Alignment for Faster Speculative Decoding

Shijing Hu, Jingyang Li, Xingyu Xie et al.

NEURIPS 2025arXiv:2502.11018
3
citations

GRIP: A Graph-Based Reasoning Instruction Producer

Jiankang Wang, Jianjun Xu, Xiaorui Wang et al.

NEURIPS 2025arXiv:2412.08864
2
citations

Grounding by Trying: LLMs with Reinforcement Learning-Enhanced Retrieval

Sheryl Hsu, Omar Khattab, Chelsea Finn et al.

ICLR 2025arXiv:2410.23214
16
citations

Guaranteed Generation from Large Language Models

Minbeom Kim, Thibaut Thonet, Jos Rozen et al.

ICLR 2025arXiv:2410.06716
5
citations

h4rm3l: A Language for Composable Jailbreak Attack Synthesis

Moussa Koulako Bala Doumbouya, Ananjan Nandi, Gabriel Poesia et al.

ICLR 2025arXiv:2408.04811
11
citations

HaDeMiF: Hallucination Detection and Mitigation in Large Language Models

Xiaoling Zhou, Mingjie Zhang, Zhemg Lee et al.

ICLR 2025
9
citations

HALL-E: Hierarchical Neural Codec Language Model for Minute-Long Zero-Shot Text-to-Speech Synthesis

Yuto Nishimura, Takumi Hirose, Masanari Ohi et al.

ICLR 2025arXiv:2410.04380
5
citations

Hansel: Output Length Controlling Framework for Large Language Models

Seoha Song, Junhyun Lee, Hyeonmok Ko

AAAI 2025paperarXiv:2412.14033
1
citations

HarmAug: Effective Data Augmentation for Knowledge Distillation of Safety Guard Models

Seanie Lee, Haebin Seong, Dong Bok Lee et al.

ICLR 2025arXiv:2410.01524
15
citations

HBLLM: Wavelet-Enhanced High-Fidelity 1-Bit Quantization for LLMs

Ningning Chen, Weicai Ye, Ying Jiang

NEURIPS 2025spotlightarXiv:2512.00862
1
citations

HC-LLM: Historical-Constrained Large Language Models for Radiology Report Generation

Tengfei Liu, Jiapu Wang, Yongli Hu et al.

AAAI 2025paperarXiv:2412.11070
8
citations

HCRMP: An LLM-Hinted Contextual Reinforcement Learning Framework for Autonomous Driving

Zhiwen Chen, Hanming Deng, Zhuoren Li et al.

NEURIPS 2025arXiv:2505.15793
3
citations

HeGTa: Leveraging Heterogeneous Graph-enhanced Large Language Models for Few-shot Complex Table Understanding

Rihui Jin, Yu Li, Guilin Qi et al.

AAAI 2025paperarXiv:2403.19723
6
citations

Herald: A Natural Language Annotated Lean 4 Dataset

Guoxiong Gao, Yutong Wang, Jiedong Jiang et al.

ICLR 2025arXiv:2410.10878
31
citations

HexGen-2: Disaggregated Generative Inference of LLMs in Heterogeneous Environment

YOUHE JIANG, Ran Yan, Binhang Yuan

ICLR 2025arXiv:2502.07903
21
citations

Hierarchical Demonstration Order Optimization for Many-shot In-Context Learning

Yinhan He, Wendy Zheng, Song Wang et al.

NEURIPS 2025

HiMoLE: Towards OOD-Robust LoRA via Hierarchical Mixture of Experts

Yinuo Jiang, Yan Xiaodong, Keyan Ding et al.

NEURIPS 2025

HMoRA: Making LLMs More Effective with Hierarchical Mixture of LoRA Experts

Mengqi Liao, Wei Chen, Junfeng Shen et al.

ICLR 2025
8
citations