LLM+P: LLMs with Classical Planners
LLM+P is a framework that combines the natural language understanding capabilities of large language models with the formal guarantees of classical AI planners.1) Introduced by Liu et al., 2023, the approach uses an LLM to translate natural language problem descriptions into Planning Domain Definition Language (PDDL), which is then solved by an established planner such as Fast Downward. This hybrid architecture leverages the strengths of both paradigms: LLMs handle ambiguous natural language input while classical planners provide optimal and correct solutions for well-defined planning problems.
The core LLM+P pipeline operates in three stages:
Problem Translation: The LLM receives a natural language task description along with a PDDL domain template. It generates the PDDL problem file specifying the initial state, goal conditions, and relevant objects.
Classical Planning: An off-the-shelf planner (e.g., Fast Downward, LAMA, FF) solves the PDDL problem, producing a guaranteed-valid action sequence.
Plan Translation: The solution is converted back to natural language or executable actions for the user or downstream agent.
This separation of concerns ensures that the LLM handles what it excels at (language understanding, common sense, disambiguation) while the planner handles what it excels at (combinatorial search with correctness guarantees).
A 2025 survey by Tantakoun, Muise, and Zhu (ACL Findings 2025) reframes LLMs not as planners themselves but as planning formalizers that construct and iteratively refine PDDL models. Key contributions:
LLMs generate initial PDDL domain and problem specifications from natural language
Feedback from the planner (e.g., unsolvable problems, invalid actions) is used to iteratively correct the PDDL formulation
This iterative loop addresses the brittleness of one-shot PDDL generation
The approach scales to long-horizon problems where pure LLM planning degrades
Integration with Classical Planners
Common planners used in LLM+P architectures:
Fast Downward (
Helmert, 2006): Supports multiple heuristic search algorithms; the most widely used in LLM+P research
2)
LAMA (Richter & Westphal, 2010): Landmark-based planner optimized for satisficing planning
FF (Hoffmann & Nebel, 2001): Fast-forward planner using relaxed plan heuristics
The 2025 International Planning Competition evaluation tested frontier LLMs (DeepSeek R1, Gemini 2.5 Pro, GPT-5) directly against LAMA on standard IPC domains. While GPT-5 was competitive on standard domains, all LLMs degraded significantly on obfuscated variants where semantic cues were removed, confirming that pure LLM planning relies heavily on pattern matching rather than formal reasoning.
Advantages Over Pure LLM Approaches
Correctness Guarantees: Classical planners produce provably valid plans when given well-formed PDDL
Optimality: Planners can find optimal or near-optimal solutions; LLMs tend to generate satisficing but suboptimal plans
Scalability: Classical planners handle large state spaces through efficient heuristic search
Interpretability: PDDL plans are human-readable formal specifications
Robustness: Plans don't suffer from hallucination or reasoning errors once correctly formalized
The LLM+P paradigm has inspired several extensions:
Limitations and Open Challenges
PDDL Coverage: Many real-world problems resist clean PDDL formulation (partial observability, continuous dynamics, stochastic effects)
Translation Errors: LLMs may generate syntactically valid but semantically incorrect PDDL, requiring iterative correction
Domain Engineering: Creating PDDL domain templates still requires planning expertise
Scalability of Translation: As problem complexity grows, accurate PDDL generation becomes harder for LLMs
Obfuscation Sensitivity: LLMs struggle when domain descriptions are unfamiliar or abstracted away from natural patterns
Active research directions include automated PDDL domain learning, end-to-end differentiable planning, and integration with reinforcement learning for problems that resist pure symbolic formulation.
See Also
References