Current methods like prior task decomposition and post-hoc experience summarization have limitations in guiding LLMs to achieve high-level goals. Prior task decomposition methods break down complex tasks into sub-tasks before environmental interaction, lacking grounded, dynamic adjustment. On the other hand, post-hoc experience summarization methods, such as Reflexion and CLIN, can produce overly detailed guidelines and may not provide suitable guidance for high-level goals6. These methods often struggle to effectively prioritize strategies and lack empirical grounding. As a result, LLMs face challenges in achieving broad, high-level goals with delayed rewards and require additional support modules for decision-making tasks.
Large language models (LLMs) face several primary challenges when tasked with broad, high-level goals. These challenges include the ambiguity of the goals, delayed rewards, and the impracticality of frequent model retraining to adapt to new goals and tasks. Additionally, current approaches for achieving high-level goals, such as prior task decomposition and post-hoc experience summarization, have limitations such as lack of empirical grounding or difficulty in effectively prioritizing strategies. These challenges make it difficult for LLMs to consistently achieve high-level goals without training while overcoming the mentioned limitations.
The Search Module within SELFGOAL enhances the selection of appropriate subgoals for specific situations by identifying the most suitable subgoals for the current situation2. It does this by selecting from the leaf nodes in the GOALTREE, which is a tree structure of textual subgoals. The module considers the current state as a description of the interaction history with the environment and regards the leaf nodes of each branch as potential subgoals2.
By using a representation of the current state, the Search Module ensures that the selected subgoals are contextually relevant and adaptable to the changing environment2. This contextual awareness allows language agents to achieve high-level goals more effectively by providing precise guidance in diverse environments.
Furthermore, the Search Module works in conjunction with the Decomposition Module, which refines the GOALTREE by breaking down selected subgoals into more concrete ones. This dynamic adjustment allows SELFGOAL to provide contextually relevant guidance and adapt to changing situations, significantly improving language agent performance in both collaborative and competitive scenarios.