"embodied agents" Papers

23 papers found

ADAM: An Embodied Causal Agent in Open-World Environments

Shu Yu, Chaochao Lu

ICLR 2025arXiv:2410.22194
10
citations

AgentSquare: Automatic LLM Agent Search in Modular Design Space

Yu Shang, Yu Li, Keyu Zhao et al.

ICLR 2025arXiv:2410.06153
61
citations

An Embodied AR Navigation Agent: Integrating BIM with Retrieval-Augmented Generation for Language Guidance

Hsuan-Kung Yang, Tsu-Ching Hsiao, Ryoichiro Oka et al.

ISMAR 2025paperarXiv:2508.16602
1
citations

EfficientNav: Towards On-Device Object-Goal Navigation with Navigation Map Caching and Retrieval

Zebin Yang, Sunjian Zheng, Tong Xie et al.

NEURIPS 2025arXiv:2510.18546

ESCA: Contextualizing Embodied Agents via Scene-Graph Generation

Jiani Huang, Amish Sethi, Matthew Kuo et al.

NEURIPS 2025oralarXiv:2510.15963

HASARD: A Benchmark for Vision-Based Safe Reinforcement Learning in Embodied Agents

Tristan Tomilin, Meng Fang, Mykola Pechenizkiy

ICLR 2025arXiv:2503.08241
5
citations

LabUtopia: High-Fidelity Simulation and Hierarchical Benchmark for Scientific Embodied Agents

Rui Li, Zixuan Hu, Wenxi Qu et al.

NEURIPS 2025arXiv:2505.22634
2
citations

Memo: Training Memory-Efficient Embodied Agents with Reinforcement Learning

Gunshi Gupta, Karmesh Yadav, Zsolt Kira et al.

NEURIPS 2025spotlightarXiv:2510.19732

MindForge: Empowering Embodied Agents with Theory of Mind for Lifelong Cultural Learning

Mircea Lică, Ojas Shirekar, Baptiste Colle et al.

NEURIPS 2025arXiv:2411.12977
1
citations

MLLM as Retriever: Interactively Learning Multimodal Retrieval for Embodied Agents

Junpeng Yue, Xinrun Xu, Börje F. Karlsson et al.

ICLR 2025arXiv:2410.03450
8
citations

Multi-Modal Grounded Planning and Efficient Replanning for Learning Embodied Agents with a Few Examples

Taewoong Kim, Byeonghwi Kim, Jonghyun Choi

AAAI 2025paperarXiv:2412.17288
7
citations

Multimodal LLM Guided Exploration and Active Mapping using Fisher Information

Wen Jiang, BOSHU LEI, Katrina Ashton et al.

ICCV 2025arXiv:2410.17422
9
citations

NeSyC: A Neuro-symbolic Continual Learner For Complex Embodied Tasks in Open Domains

Wonje Choi, Jinwoo Park, Sanghyun Ahn et al.

ICLR 2025arXiv:2503.00870
8
citations

Provable Ordering and Continuity in Vision-Language Pretraining for Generalizable Embodied Agents

Zhizhen Zhang, Lei Zhu, Zhen Fang et al.

NEURIPS 2025oralarXiv:2502.01218
2
citations

SEEA-R1: Tree-Structured Reinforcement Fine-Tuning for Self-Evolving Embodied Agents

Wanxin Tian, Shijie Zhang, Kevin Zhang et al.

NEURIPS 2025arXiv:2506.21669
6
citations

ViSPLA: Visual Iterative Self-Prompting for Language-Guided 3D Affordance Learning

Hritam Basak, Zhaozheng Yin

NEURIPS 2025

Visual Agentic AI for Spatial Reasoning with a Dynamic API

Damiano Marsili, Rohun Agrawal, Yisong Yue et al.

CVPR 2025arXiv:2502.06787
31
citations

Fast-Slow Test-Time Adaptation for Online Vision-and-Language Navigation

JUNYU GAO, Xuan Yao, Changsheng Xu

ICML 2024arXiv:2311.13209
17
citations

Improving Knowledge Extraction from LLMs for Task Learning through Agent Analysis

James Kirk, Robert Wray, Peter Lindes et al.

AAAI 2024paperarXiv:2306.06770
7
citations

NavGPT: Explicit Reasoning in Vision-and-Language Navigation with Large Language Models

Gengze Zhou, Yicong Hong, Qi Wu

AAAI 2024paperarXiv:2305.16986
283
citations

Retrieval-Augmented Embodied Agents

Yichen Zhu, Zhicai Ou, Xiaofeng Mou et al.

CVPR 2024arXiv:2404.11699
28
citations

RoboMP$^2$: A Robotic Multimodal Perception-Planning Framework with Multimodal Large Language Models

Qi Lv, Hao Li, Xiang Deng et al.

ICML 2024arXiv:2404.04929
4
citations

VEON: Vocabulary-Enhanced Occupancy Prediction

Jilai Zheng, Pin Tang, Zhongdao Wang et al.

ECCV 2024arXiv:2407.12294
15
citations