跳到主要导航 跳到搜索 跳到主要内容

LLM-Guided Reinforcement Learning for Interactive Environments

  • Fuxue Yang
  • , Jiawen Liu
  • , Kan Li*
  • *此作品的通讯作者
  • Beijing Institute of Technology

科研成果: 期刊稿件文章同行评审

摘要

We propose herein LLM-Guided Reinforcement Learning (LGRL), a novel framework that leverages large language models (LLMs) to decompose high-level objectives into a sequence of manageable subgoals in interactive environments. Our approach decouples high-level planning from low-level action execution by dynamically generating context-aware subgoals that guide the reinforcement learning (RL) agent. During training, intermediate subgoals—each associated with partial rewards—are generated based on the agent’s current progress, providing fine-grained feedback that facilitates structured exploration and accelerates convergence. At inference, a chain-of-thought strategy is employed, enabling the LLM to adaptively update subgoals in response to evolving environmental states. Although demonstrated on a representative interactive setting, our method is generalizable to a wide range of complex, goal-oriented tasks. Experimental results show that LGRL achieves higher success rates, improved efficiency, and faster convergence compared to baseline approaches.

源语言英语
文章编号1932
期刊Mathematics
13
12
DOI
出版状态已出版 - 6月 2025
已对外发布

指纹

探究 'LLM-Guided Reinforcement Learning for Interactive Environments' 的科研主题。它们共同构成独一无二的指纹。

引用此