Mastering Long-Horizon Planning with GRASP: A Q&A Guide
This article explores GRASP, a novel gradient-based planner designed for learned world models. As predictive models become more powerful, they promise general-purpose simulation, but long-horizon planning remains fragile due to ill-conditioned optimization, poor local minima, and high-dimensional visual spaces. GRASP introduces three key innovations—virtual states for parallel optimization, direct state stochasticity for exploration, and gradient reshaping to bypass brittle visual gradients—to make planning practical over extended horizons. Below, we answer common questions about world models, the challenges of long-horizon planning, and how GRASP overcomes them.
What is a world model, as defined in this research?
The term "world model" is often overloaded. In some contexts, it means an explicit dynamics model that predicts future states; in others, it refers to an implicit internal state used by generative models (e.g., an LLM playing chess). For GRASP, we adopt a practical definition: a learned model that, given the current state and a sequence of future actions, predicts what happens next. Formally, it approximates the environment's transition dynamics by defining a predictive distribution Pθ(st+1 | st-h:t, at). This model can work on high-dimensional observations like images, latent vectors, or proprioception, enabling it to function as a general-purpose simulator across tasks.

Why is long-horizon planning with world models particularly challenging?
Planning over extended horizons introduces several failure modes. First, optimization becomes ill-conditioned because small changes in early actions can have exponentially amplified effects later, making gradient signals noisy or vanishing. Second, non-greedy structure—where the best short-term action leads to a dead end—creates poor local minima that trap optimizers. Third, high-dimensional latent spaces in modern vision-based world models are rife with subtle error surfaces, so gradient-based planners often wander into unrealistic states or get stuck. These issues compound as the planning horizon lengthens, turning what seems like a straightforward optimization into a brittle process that fails unpredictably.
What is GRASP, and what are its three core innovations?
GRASP (Gradient-based RAndom SPlitting) is a planner designed to make gradient-based optimization in learned world models robust over long horizons. It introduces three key ideas:
- Virtual states: Instead of optimizing actions sequentially, GRASP lifts the trajectory into a set of virtual states, allowing parallel optimization across all time steps.
- Stochastic state iterates: It adds controlled noise directly to the state predictions during optimization, helping the planner explore alternative trajectories and escape local minima.
- Gradient reshaping: It modifies the gradient flow so that action signals remain clean and informative, avoiding the brittle gradients that pass through high-dimensional vision encoders.
Together, these innovations make long-horizon planning practical where previous gradient-based methods would fail.
How does GRASP use virtual states to parallelize optimization across time?
Classic gradient-based planning unfolds a trajectory step by step: at each time, the current state is used to compute the next state via the world model, and gradients are backpropagated through the entire chain. This sequential dependency limits parallelism and can cause vanishing gradients over many steps. GRASP instead introduces virtual states—intermediate optimization variables that represent the state at each time step but are not directly tied to the previous state. The planner then optimizes these virtual states in parallel, using a consistency constraint to ensure they correspond to a valid trajectory. This decoupling allows the algorithm to compute gradients for all time steps simultaneously, drastically speeding up optimization and improving gradient flow.

Why does GRASP add stochasticity directly to state iterates, and how does it help exploration?
In long-horizon planning, deterministic optimization can get stuck in narrow valleys or ignore promising but non-obvious action sequences. GRASP addresses this by injecting controlled stochastic noise into the state predictions during the planning process. This is not just random exploration in action space; it perturbs the latent state iterates themselves, allowing the optimizer to sample different trajectories. The noise helps the planner escape poor local minima and explore diverse futures without relying on expensive random restarts or brute-force sampling. By carefully tuning the variance and annealing it over iterations, GRASP balances exploration and exploitation, making the optimization more robust over long horizons.
How does GRASP reshape gradients to avoid issues with high-dimensional vision models?
World models that use high-dimensional visual encoders (like CNNs or transformers) produce "state-input gradients" that are notoriously brittle: they often vanish, explode, or carry little meaningful information about actions. GRASP solves this by reshaping the gradient flow so that action selection receives clean, direct signals. Instead of backpropagating through the entire vision model to update actions, GRASP computes a modified gradient that bypasses the visual encoder's idiosyncrasies. It does this by treating the world model's latent dynamics separately from the observation encoder, and by using the virtual state formulation to decouple action updates from pixel-level gradients. The result is that the planner can more reliably optimize actions even when the world model operates in a high-dimensional visual space, without getting lost in noisy gradients.