Portrait of Fuyang (Scott) Cui

Fuyang (Scott) Cui

Computer Science @ University of Toronto

Dec 1, 2025

CausalPool: Do VLMs Learn Physics or Just Match Patterns?

Investigating the 'Causal Disconnect' in Vision-Language Models through counterfactual supervision.

Deep LearningCausalityVLMsResearch

CausalPool: Eliciting Causal Behaviour through Counterfactual Supervision

Paper & Resources:


Do Vision-Language Models (VLMs) actually construct internal world models of physical dynamics? Or are they merely sophisticated “causal parrots”?

While a model might correctly predict that a glass will shatter if dropped, does it understand the why and how, or is it just correlating images of falling objects with broken shards?

In our paper CausalPool, we investigate this by testing a fundamental hypothesis: If a model can reason counterfactually (simulating “what if”), does it necessarily understand the physical reality (“what is”)?

Our results uncover a surprising Causal Disconnect.

🪜 The Ladder of Causation

To distinguish between correlation and true understanding, we use Judea Pearl’s Ladder of Causation.

Figure 1: Overview of the CausalPool Dataset. Top: Key frames from a sample video. Bottom: Examples of the four distinct question types generated for each simulation. Figure 1: Overview of the CausalPool Dataset. We test models on Descriptive queries (Rung 1) and Counterfactual queries (Rung 3) involving continuous physics variables.

  1. Rung 1: Association (Observation)
    • “What happened in this video?”
    • Requires robust perception.
  2. Rung 2: Intervention (Action)
    • “What if I do X?”
  3. Rung 3: Counterfactuals (Imagination)
    • “What would have happened if the ball started faster?”
    • This is the highest form of reasoning. It requires abduction: inferring the hidden state of the world from observation, then modifying it.

The Top-Down Hypothesis

Intuitively, Counterfactual reasoning should be the hardest task. To simulate a deviation from reality, you must first correctly perceive reality. Therefore, training a model to master Rung 3 should force it to master Rung 1.

We set out to test this top-down learning curriculum.

🎱 Introducing CausalPool

We introduce CausalPool, a synthetic video benchmark generated via the Pooltool physics engine. Unlike previous benchmarks that rely on discrete collisions, CausalPool focuses on continuous variable interactions (velocity vectors, friction, angular momentum).

Figure 2: Overview of data generating process. Pool Tool Simulations generate videos and metadata, which are parsed to create descriptive, predictive, and counterfactual query pairs. Figure 2: Data Generation Pipeline. We generate 16,384 video clips with precise metadata. We then synthesize questions that blind the model to “what is” while quizzing it on “what if.”

  • Total Clips: 16,384 (single-ball scenarios)
  • Interventions: Precise changes to Position (x,yx, y) and Velocity (dx,dydx, dy).
  • The Experiment: We trained specialized models (Qwen3-VL-4B) exclusively on specific rungs of the ladder to see if capabilities transfer.

📉 The Causal Disconnect

Our experiments revealed a counter-intuitive result that challenges the “Top-Down” hypothesis.

We trained a model exclusively on Counterfactual queries (“If speed doubled, what happens?”). We expected this model to learn the underlying physics and thus be able to describe the original video.

It failed.

ModelTraining DataRung 3 Accuracy (Counterfactual)Rung 1 Accuracy (Descriptive)
Random Chance-25%25%
Zero-Shot BaselineNone~27%~38%
CausalPool-DescDescriptive Only38%84%
CausalPool-CFCounterfactual Only70%45%

Key Findings:

  1. The Specialist: The model trained on counterfactuals (CausalPool-CF) became excellent at counterfactuals (70% accuracy).
  2. The Failure: That same model showed negligible improvement on describing the video (45%) compared to the baseline.
  3. The Conclusion: The model learned to be a “Blind Simulator.”

It learned a function to map linguistic interventions to outcomes via shortcut mechanisms, without grounding these predictions in the visual reality of the video.

👁️ The “Blind Simulator” Phenomenon

The failure is best illustrated qualitatively. In the example below, the model correctly answers a complex math-heavy counterfactual question but hallucinates events in the actual video.

Figure 3: Contrast between Descriptive Failure and Counterfactual Success. (Left) The model hallucinates a wall collision. (Right) The model correctly predicts a trajectory change. Figure 3: The Causal Disconnect. (Left) When asked “What happened?”, the model hallucinates a wall collision (Option D). (Right) Yet, when asked to simulate a hypothetical trajectory where the velocity is inverted, the model correctly identifies the outcome (Option C).

  • Rung 1 (Reality): The model hallucinates a wall collision that never happened.
  • Rung 3 (Imagination): The model correctly simulates a complex trajectory change.

This suggests the model solves the reasoning task by bypassing the visual encoder in favor of learning language-based transition dynamics. It solves the math, but it ignores the pixels.

Future Directions

Our work demonstrates that Supervised Fine-Tuning (SFT) is insufficient to induce robust physical world models. Current architectures can decouple logical reasoning from visual perception. To bridge this gap, future work must move beyond next-token prediction toward Causal Reinforcement Learning or neuro-symbolic grounding.