LLMs and World Models: Adding a planet destroys the system


Ben was the one running the experiment. I was mostly watching and listening. He had built an AI “coach” and trained it very carefully. The rule was simple. The coach was not allowed to give answers. It was only supposed to ask questions, like a good teacher does. The goal was to help someone think, not to think for them.

Then Ben fed it a research paper about world models. The paper talked about how AI systems learn patterns from data generated by real physical laws, like Newton’s or Kepler’s laws of motion. It’s heavy stuff. Even for humans, it takes work to understand.

At first, the AI coach behaved. It asked Ben to explain the paper in his own words. It waited. It nudged him gently. It looked like teaching.

Then, slowly, it broke character.

Not all at once. Just a little slip. A helpful sentence here. A partial answer there. Over time, the coach stopped being a coach. It turned into what these systems always turn into: an answer machine.

That moment mattered more to me than anything else in the paper.

Because what we were seeing wasn’t a bug. It was a feature.

Ben explained the paper’s main finding. Imagine you train a model on data from a single planet orbiting a sun. The model learns that orbit perfectly. It looks amazing. But then you add a second planet. Suddenly, the model falls apart.

A human who understands gravity wouldn’t panic. Gravity doesn’t care how many planets there are. But the model didn’t learn gravity. It learned one very good trick.

It memorized a curve.

The paper showed this clearly. In the real Newtonian model, the force vectors all point cleanly toward the sun. In the AI’s learned model, the vectors are crooked and uneven. The math is fancy. The results look fine, until they don’t.

The model didn’t understand the world. It found something that worked and stopped there.

That’s when I jumped in.

I told Ben I see this all the time in coding. People call it “vibe coding.” You ask an AI to build a website. It does it fast. The page loads. The buttons work. Everyone claps.

Then you open the code.

It’s a mess.

Files everywhere. Logic copied and pasted three times. Hacks stacked on top of hacks. It’s like a bridge made of cardboard and duct tape. You can walk across it once. Maybe twice. But don’t add traffic.

The AI didn’t reason its way to a clean design. It didn’t ask, “What’s the right structure?” It asked, “What usually works?”

And that’s the key problem.

Large language models are not reasoners. They are pattern machines. Very good ones. But still pattern machines.

They don’t hold ideas steady. They don’t protect intent. The AI coach didn’t “decide” to stop asking questions. It drifted there because giving answers is the safest pattern. Over a long conversation, safety beats rules.

This is what I call conversational integrity failure. The longer the exchange goes on, the harder it is for the model to stay true to its role. Everything becomes tokens. Instructions become suggestions. Purpose melts away.

On the surface, it still sounds smart. It still explains things. It still feels helpful. But under the hood, there is no world model holding it together.

No gravity.

That’s why these systems feel magical right up until they don’t. They can talk about thinking. They can fake reasoning. They can even explain their own “logic.” But those explanations are just more patterns.

They are stories told after the fact.

This conversation with Professor Ben Teehankee was part of a much longer discussion about AI and the future of learning. We weren’t trying to hype AI or tear it down. We were trying to understand something quieter and more unsettling: why machines that sound so smart can still miss the point so badly.

Ben and I weren’t angry about this. We weren’t scared either. But we were very clear-eyed. These tools are powerful, but they are shallow in a specific way. They optimize for the next step, not the big picture. They chase results, not truth.

If we treat them like minds, we will keep being disappointed. If we treat them like tools that guess very well, we’ll do better.

Because the moment you add a new planet, or ask for a longer conversation, the cracks show. And once you see those cracks, you can’t unsee them.

 

About Me:

Dominic “Doc” Ligot is one of the leading voices in AI in the Philippines. Doc has been extensively cited in local and global media outlets including The Economist, South China Morning Post, Washington Post, and Agence France Presse. His award-winning work has been recognized and published by prestigious organizations such as NASA, Data.org, Digital Public Goods Alliance, the Group on Earth Observations (GEO), the United Nations Development Programme (UNDP), the World Health Organization (WHO), and UNICEF.

If you need guidance or training in maximizing AI for your career or business, reach out to Doc via https://docligot.com.

Follow Doc Ligot on Facebook: https://facebook.com/docligotAI