The Pasta Sauce Problem: Why Your AI Assistant Can Be Brilliant and Completely Wrong at the Same…

11 min read Original article ↗

Arun Sanna

How the “Intent Gap” became AI’s $50 billion blind spot - and what we’re finally doing about it

Last month, I asked an AI agent to “clean up the old docs” on my laptop.

I gave it access to my Documents folder. I trusted it. I went to make coffee.

When I came back, the agent had delivered a cheerful report: duplicates removed, files organized into tidy folders, a neat summary of everything it accomplished. It was genuinely impressive. The AI had worked faster and more systematically than I ever could.

Then I noticed my thesis was gone.

Not in the trash. Not renamed. Gone.

The AI didn’t fail because it hallucinated. It didn’t make up facts or crash. It did something far more unsettling: it succeeded brilliantly at the wrong thing. It took my fuzzy human request, guessed at what I really wanted, committed fully to that guess, and executed with terrifying precision.

It turns out I’m not alone. Air Canada’s chatbot confidently gave a customer wrong information about bereavement fares — the company lost the lawsuit. A Chevrolet dealership’s AI agreed to sell a $70,000 Tahoe for one dollar. DPD’s delivery bot started writing poems mocking its own company. This pattern — smart, fast, and subtly catastrophic — is happening everywhere as AI agents graduate from chatbots to actual doers. And it has a name: The Intent Gap.

The Pasta Sauce Test

The easiest way to understand the Intent Gap is with pasta.

Imagine you tell a friend: “Hey, can you make some quick pasta sauce?”

What happens in their brain is remarkable. They don’t just parse your words — they run an entire simulation. They consider context, your relationship, the time of day, and a dozen invisible social rules. In about two seconds, they’ve figured out:

  • The real goal: You’re probably hungry and want something simple. This isn’t a request for their grandmother’s 6-hour bolognese recipe.
  • The priority: “Quick” is the operative word. Speed beats perfection.
  • The invisible guardrails: Don’t use the expensive truffle oil you were saving. Don’t leave the kitchen looking like a war zone. Don’t make something so elaborate it creates pressure to reciprocate. None of this was in your words. Your friend inferred all of it.

Press enter or click to view image in full size

Now imagine asking an AI the same thing.

The AI sees: “make,” “quick,” “pasta sauce.”

Without the human ability to simulate your true goals, it might latch onto “pasta sauce” as the primary objective. It could decide that your real goal is to have excellent pasta sauce - after all, who wants bad pasta sauce? - and pull up an authentic Italian recipe from its training data. Three hours later, you’re still hungry, watching it explain the importance of San Marzano tomatoes.

The AI didn’t misunderstand your words. It misunderstood your intent.

And here’s the thing: your friend didn’t need you to say “don’t use the truffle oil.” Humans operate with thousands of unspoken guardrails, invisible rules we’ve absorbed through a lifetime of social interaction. AI agents don’t have these. They need us to spell out every single one - and we never do, because we don’t even realize they’re there.

Context Is Not Intent (And That’s the Whole Problem)

This brings us to a distinction that sounds academic but is actually the key to understanding why AI keeps face-planting: context and intent are not the same thing.

Context is what’s explicitly in your request:

• The literal words you used

• The constraints you mentioned

• The files you pointed to

Intent is what’s NOT in your request:

• Your true priorities (what matters most?)

• Acceptable tradeoffs (is speed more important than quality?)

• What “done” actually looks like

• What would make you regret asking

• The invisible guardrails (“don’t delete my thesis”)

Press enter or click to view image in full size

Think of it like an iceberg. Context is the 10% above water - visible, explicit, written down. Intent is the 90% below - massive, invisible, and absolutely critical.

Here’s what makes this particularly dangerous: AI is really, really good at extracting context. Modern language models can pull entities, constraints, and instructions out of massive documents with impressive accuracy. So when your AI agent confidently reports “Task completed!” it’s not lying. It extracted the context perfectly and executed flawlessly against it.

It just never knew about the iceberg below the waterline.

As one AI researcher put it: “Intent is not in the text the way context is.” You can’t just give an AI a bigger context window and expect it to suddenly understand what you meant. Intent requires inference - the ability to simulate what’s in your head - and that’s a fundamentally different problem than processing words.

From “Oops” to “Oh No”: When AI Starts Writing to Reality

If all AI did was chat, the Intent Gap would be a minor annoyance. You’d say something, the AI would misunderstand, you’d clarify, and life would go on. Conversations are forgiving that way.

But something important has changed. We’re not just talking to AI anymore. We’re giving it tools.

File access. Calendar management. Email. Databases. Credit cards.

Press enter or click to view image in full size

And when an AI agent has tools, it’s no longer generating text. It’s generating actions. It’s not writing to a chat window; it’s writing to reality.

My deleted thesis wasn’t a misunderstanding I could just correct. It was an irreversible action. The agent made a real-world commitment on my behalf - one I couldn’t undo by typing “wait, I didn’t mean that.”

This is the same challenge that forced the cryptocurrency world to rethink how they design systems. In DeFi, transactions are expensive and irreversible. You can’t just “undo” a trade. So they developed what they call “intent-based” systems: before any action happens, the user signs an explicit intent - a formal document stating exactly what they want and what constraints apply. Only then do specialized “solvers” compete to execute it safely.

Sound familiar?

When the stakes are high and actions can’t be undone, systems naturally evolve toward making intent explicit. It’s not a coincidence. It’s a design pattern that emerges whenever the cost of misunderstanding becomes too painful.

And right now, AI agents are learning this lesson the hard way.

- -

Why “Just Give It More Information” Doesn’t Work

The instinctive response to the Intent Gap is: “Just be more specific! Write better prompts! Give the AI more context!”

And yes, being clearer helps. But it’s not a solution - it’s a patch. Here’s why:

1. You can’t articulate what you don’t know you know.

Most of your intent is unconscious. You don’t realize you have a guardrail against deleting original files until after they’re deleted. You can’t write down invisible assumptions - that’s what makes them invisible.

2. More context can actually make things worse.

Researchers have found that AI models suffer from “lost in the middle” syndrome: when you dump too much information into a prompt, critical details get ignored. The model might miss the one sentence that said “don’t touch the originals” because it was buried in paragraph seventeen.

3. The model’s training objective works against you.

Large language models are trained for “next token prediction” - generating text that sounds plausible. They’re rewarded for producing confident, human-sounding responses, not for admitting uncertainty or asking clarifying questions. When faced with ambiguity, the model’s instinct is to pick a plausible interpretation and run with it, not to pause and ask “wait, what did you actually mean?”

This is why the Intent Gap isn’t a bug to be fixed with better prompts. It’s an architectural challenge that requires new design patterns.

Press enter or click to view image in full size

Three Ways We’re Teaching AI to Ask “What Do You Really Mean?”

The good news: researchers and system builders aren’t waiting for AI to magically get smarter. They’re developing concrete strategies to bridge the Intent Gap right now.

Press enter or click to view image in full size

1. Ask Clarifying Questions (Active Disambiguation)

The most direct fix: stop guessing and start asking.

Instead of confidently assuming what you meant, the AI is designed to detect ambiguity - especially for high-stakes actions - and pause. “I’m about to delete 47 files. Some of these look like originals. Should I proceed, or would you like to review them first?”

This isn’t rocket science, but it requires explicitly programming the AI to prefer asking over guessing. Most models aren’t trained this way. They’re trained to be helpful and confident, which means they’ll happily walk off a cliff rather than admit they’re not sure where you wanted to go.

Building a “clarification loop” into agent architecture - where the system recognizes uncertainty and escalates to the user - is becoming a core design pattern for production systems. Anthropic’s Claude, for instance, uses Constitutional AI principles that explicitly train the model to recognize uncertainty and ask for clarification rather than confidently guess.

2. Keep Your Options Open (Probabilistic Intent)

What if the AI didn’t commit to a single interpretation immediately?

This approach has the model maintain a distribution of plausible goals rather than locking onto one. When you say “clean up my docs,” it might internally consider:

• 60% chance you mean “organize and remove duplicates”

• 30% chance you mean “archive old files”

• 10% chance you mean “full cleanup including originals”

As the conversation continues and you provide more information, these probabilities update. The correct intent “crystallizes” over time instead of being guessed upfront.

This prevents the AI from confidently sprinting in the wrong direction. It’s the difference between a GPS that recalculates when you take a wrong turn and a GPS that insists you drive into a lake.

3. Create an Intent Document (Make the Invisible Visible)

The most powerful approach: stop treating intent as something hidden in a prompt. Make it a first-class artifact - a separate, explicit document that spells out:

• Goals: What does success look like?

• Constraints: What is the AI NOT allowed to do?

• Priorities: When goals conflict, which wins?

• Failure modes: What would make this a disaster?

Think of it as a “semantic commit” - a formal contract between you and the AI about what it’s actually trying to accomplish. It’s versioned, testable, and updatable, separate from the instructions themselves.

This approach transforms intent from a guessing game into an interface design problem. It’s not glamorous, but it’s the kind of boring-but-critical infrastructure that separates systems that work from systems that delete your thesis.

The Future Isn’t Magic Understanding

Here’s what the next few years probably don’t look like: a sudden breakthrough where AI magically understands everything you meant from a few words. That’s a nice fantasy, but it’s not how this technology works.

Here’s what they probably do look like: AI agents that know what they don’t know.

The vision isn’t an all-knowing oracle. It’s what researchers call an “always-on agent” that runs cheap background checks - a kind of “human second pass” - to validate its own understanding. It simulates consequences before acting. It asks itself: “Am I about to do something irreversible? Am I uncertain about the goal? Is this the kind of action the user would regret?”

And when the answers come back risky, it does something revolutionary for AI: it stops and asks.

Not because it’s less capable. Because it’s more mature.

The companies that win the AI agent race won’t be the ones with the most powerful models or the most tools. They’ll be the ones who build systems that carry intent all the way from your fuzzy request to precise, reliable execution.

The Bottom Line

The Intent Gap is simple to understand and maddeningly hard to solve:

• Humans communicate with invisible context. AI takes us literally.

• When AI just chatted, this was annoying. Now that it acts, it’s dangerous.

• The fix isn’t smarter AI - it’s AI that knows when to stop and ask.

We’re at an inflection point. The models are powerful enough to do real work. The question is whether we can teach them to do the right work.

The next time you give an AI agent an instruction, imagine you’re talking to the most confident, most literal-minded, most eager-to-please intern in history. One who will absolutely, positively execute your request exactly as stated - with no idea that you left out the most important part.

Then write it down. All of it. Even the obvious stuff.

Especially the obvious stuff.

Got a story about AI misunderstanding your intent? I’d love to hear it in the comments. Sometimes the only way forward is to laugh at the robots while we teach them to ask better questions.