Engineering

I Stopped Whispering to AI (And Started Producing It)

Context engineering isn't about prompting harder—it's about producing smarter.

Ofer Avnery By Ofer Avnery
@ · 10/1/25

TL;DR: Stop treating AI like a magic spell that needs the perfect incantation. Start treating it like a live broadcast that needs a great production team.

I'll never forget the moment it clicked.

I was sitting in a meeting, watching a colleague spend thirty minutes crafting what he called the "perfect prompt." He was tweaking every word, adding examples, even begging the AI to "think step by step" like he was casting some kind of spell.

Meanwhile, across the hall? Chaos. The sales team's AI was quoting prices that made no sense. Support's chatbot was confidently citing "sources" that didn't exist. And everyone—everyone—kept saying the same thing:

"We just need better prompts."

That's when it hit me: We've been thinking about AI completely backwards.

The Lie We're All Believing

Here's what every AI vendor wants you to think: "Our next model will be so smart, you can just dump everything into it and—boom—magic happens."

Google announces they can handle 2 million tokens (that's like 3,000 pages of text). Anthropic shows off AI agents that can work for hours without supervision. The headlines won't stop screaming about "reasoning breakthroughs."

And look—they're not lying. These advances are genuinely impressive.

But Here's What They Don't Say

None of it works without someone directing the show. You can have the smartest AI in the world, but if you're just throwing information at it and hoping for the best? You're setting yourself up for expensive failures.

What CNN Knows That Most Businesses Don't

Let me ask you something: Have you ever watched Anderson Cooper during breaking news? He's calm, collected, delivering exactly the right information at exactly the right moment.

Do you think he's just... really good at memorizing everything that's ever happened?

Of course not. That would be impossible.

Here's what's actually happening: Behind Anderson Cooper is an entire team working like crazy. Producers are routing news stories from all over the world. Fact-checkers are verifying every claim. Researchers are digging up relevant footage. And someone's constantly feeding him what to say next, exactly when he needs it.

Anderson Cooper isn't winging it. He's being produced by a well-oiled machine.

That's what context engineering really is. Your AI is Anderson Cooper. The question is: do you have a production team backing it up?

The Four Things Every Great News Show Has (That Your AI Desperately Needs)

1. A Clear Script (Not Crossed Fingers)

Before every broadcast, the anchor gets a one-page brief. It's crystal clear: What's today's story? Who are we talking to? What topics are off-limits?

Your AI needs exactly the same thing. But here's what most teams actually do: They type "be helpful" in a text box and pray it works out.

Here's what worked for us: We created an actual instruction sheet.

Not a 500-word essay. Just clear directives: "Your job is to convert trial users to paid customers. Never give legal advice. Always include: your answer, where you got it from, and what to do next."

The moment we did this? Everything changed. The AI stopped being confused. Our systems stopped breaking. We went from begging the AI to cooperate to actually directing it.

Think of it this way: You wouldn't hire a salesperson and just say "be helpful." You'd give them training, scripts, and boundaries. Your AI needs the same.

2. A Support Team (Not Magic Powers)

When Anderson Cooper says "our correspondent in London reports..." he's not making that up on the spot. Someone just fed him that information through his earpiece, seconds ago.

Your AI needs the same kind of backup. We call them "tools," but really, they're specialized helpers doing specific jobs:

  • SearchDocs – looks up information in your documentation
  • GetCRM – pulls customer data when needed
  • CreateTicket – files a task for follow-up

Each one does exactly one thing, does it well, and gets out of the way.

Here's What Nobody Tells You

Giving your AI fifty different tools is like having fifty people shouting in someone's ear at the same time. It's chaos.

We learned to keep it simple: Five tools maximum. Each with a clear name and clear purpose. The result? Our AI became faster and more accurate because it wasn't drowning in options.

3. Curated Content (Not Information Overload)

You know what news producers don't do? They don't show the anchor ten hours of raw footage and say "figure it out yourself."

Instead, they carefully select the perfect 30-second clip that tells the story.

But here's what we keep doing with AI: We dump entire 100-page documents into it and then wonder why it makes stuff up or misses key details.

The fix is simpler than you think: Only give the AI what it needs, exactly when it needs it.

When a customer asks about pricing, we don't feed the AI your entire product catalog. We grab the three relevant pricing options and only those three. Then we make sure the most important one shows up first.

Even though AI can technically handle millions of words, we've found that being selective beats being exhaustive. Every. Single. Time.

Think of it like this: You wouldn't CC someone on every email in your company and expect them to find the one relevant message. So why do that to your AI?

4. Regular Recaps (Not Endless Marathons)

Here's something the AI companies don't love to advertise: Even with those massive "context windows" everyone brags about, AI models still lose focus. They drift. They forget what you said ten messages ago.

It's like when you're scrolling through a super long Slack thread and you honestly can't remember how the conversation even started. AI has the same problem—researchers even gave it a name: "lost in the middle syndrome."

But you know who figured this out decades ago? News producers. Their solution: Recap between segments.

We do the same thing with our AI conversations.

Every few exchanges, we create a quick 6-10 line summary of what's happened so far. When someone comes back hours (or days) later, we show the AI that recap—not the entire 50-message conversation history.

It's not fancy. It's just good organizational discipline. And it absolutely works.

Think about returning to a meeting after lunch. You don't want someone to read every single thing that was said. You want: "Here's what we decided, here's where we're stuck, here's what we're doing next."

A Real Example: From $300 Leads to $3,000 Deals

Let me show you what this looks like in practice. We worked with Canada's third-largest cleaning company. Here was their problem:

The Pain: Customers would ask "How much to clean our 5,000 sq ft office twice a week?" Their chatbot would say "Someone will call you back." Meanwhile, competitors with instant quotes were closing the deals.

The Scale: 470+ service combinations (47 base services × variations for frequency, add-ons, seasonal adjustments). Try teaching an AI to quote accurately across all that.

First Attempts: Everything We Tried That Failed

Massive prompts: Explained all 470+ variations. AI quoted $1,200 for $3,400 jobs.
Vector database KB: Tried semantic search across pricing docs. Too slow, wrong matches.
RAG with embeddings: Retrieved irrelevant chunks, hallucinated combinations.

Sales spent more time fixing mistakes than closing deals.

The Fix: We stopped feeding the AI everything and started producing answers. Five specialized tools:

Get Pricing
Fetches relevant rates
Calculate Discount
Applies volume pricing
Check Availability
Verifies service area
Build Cart
Assembles the quote
Create Order
Closes the deal

The Process: Natural conversation → Pull 3 relevant options (not all 47) → Verify prices in real-time → Suggest smart upsells → Instant checkout → Hand complex cases to humans with full context.

The Results

Before: Lead capture tool generating $300 quotes
After: Revenue machine closing $3,000+ contracts 24/7

✓ Zero made-up prices
✓ Every quote backed by real spreadsheet data
✓ Sales team focuses on high-value deals, not data entry

"Can we build other services this way?" — VP of Sales

What Changed in 2025 (And What Stayed Exactly the Same)

Look, the technology really has gotten impressive. The models are bigger. They reason better. Those million-token context windows? They're real.

But here's what the benchmarks quietly showed us (and what vendors don't love to highlight):

  • AI still misses stuff buried in long documents – Just because it can read 1,000 pages doesn't mean it will catch that one crucial detail on page 847
  • Complexity hurts performance more than length – A simple 100-page document works better than a messy 20-page one
  • The best results come from AI that can look things up AND think – Not just AI that tries to remember everything

Box's CEO nailed it recently: The future isn't one super-smart AI with everything loaded into its brain. It's specialized AI agents, each with exactly the right information for their specific job.

Just like a newsroom splits up the work. The teams winning right now aren't the ones with the biggest context windows. They're the ones with the clearest direction.

The Mindset Shift That Actually Matters

Stop thinking in prompts. Start thinking in production:

  • Don't overload the system. Give it what it needs, when it needs it. (Those huge context windows are great, but curation still wins.)
  • Don't make one AI do everything. Build specialized helpers with clear roles. (That's what those "AI tools" really are—a support team.)
  • Don't guess what went wrong. Review what happened, fix your instructions, update your tools, add better examples. (Continuous improvement, not one-and-done prompting.)

If You Take Away Just One Thing

Great AI isn't about finding the perfect words.

Great AI is about great production.

The teams that master this won't look like wizards casting spells. They'll look like confident professionals running a tight operation—calm under pressure, accurate with facts, consistent day after day after day.

Ask Yourself This

Next time you're staring at a prompt box, trying to word something just right...

Ask yourself: Am I trying to find the magic words?

Or am I thinking about how to produce the right outcome?

What's one thing you're currently "prompting" that you could start "producing" instead?

The shift from prompt-whispering to production thinking isn't just about better AI outputs—it's about building systems that scale, creating teams that collaborate effectively, and delivering results that actually last.

Because at the end of the day, you don't need better spells. You need better production.

Back to Real Wave Blog