When Agentic AI Takes the Wheel, Who’s Watching the Road?
By Asaf Shamly | September 4, 2025

In advertising, speed has always had a kind of magic to it.
Campaigns launched overnight. Creatives swapped in real time. Optimizations running on a loop. The faster we moved, the smarter it felt.
Now, we’re seeing a new kind of speed enter the conversation: Agentic AI planning, selecting, and filtering autonomously. Seemingly? Decision-making on steroids.
But there’s a catch we’re starting to overlook.
As we hand more power to machines, we rarely stop to ask what those systems are learning from. What signals are they trained on? What realities are they exposed to?
Because the answer.. isn’t always reassuring.
Most of what we call AI today is trained in the shadows.
Across media buying, we see “intelligent” systems optimize for outcomes based on what they’re allowed to see: categories, averages, keyword labels, cost thresholds.
That’s not intelligence. That’s a spreadsheet with ambitions.
These inputs weren’t designed to reflect human behavior. They were built to file content, apply filters, and keep buying cycles moving. When we promote them to the role of navigators – agents acting on our behalf – we don’t get insight. We get fast decisions made inside a narrow, abstracted frame.
These systems don’t see your audience. They see your content’s metadata.
A placement is considered “safe” because it passed a keyword filter.
An impression is “viewable” because it met the minimum pixel threshold.
A site is “premium” because it sits on the right domain list.
None of this reflects what actually happens when a person lands on a page.
These systems don’t see how a user scrolls or how fast. They don’t notice that three ads are stacked above the fold, creating fatigue before the page even loads. They don’t track that the tone of the article clashes with the ad creative. They can’t sense that a comment thread is quietly hijacking brand perception.
They treat each ad like an isolated unit, when attention, emotion, and relevance build or collapse, across time and space.
And because these systems rarely receive feedback from what happened after the impression, they don’t evolve.
They repeat.
Data without depth isn’t a shortcut. It’s a liability.
The assumption about autonomous decision-making today is this: if we move fast enough through as many signals as we can, the noise will cancel itself out.
But in reality, surface-level data creates surface-level outcomes.
And when everyone is using the same incomplete inputs, it becomes harder – not easier – to compete.
This is where the real risk sits for brands who trust machines to make expert level, nuanced choices, with context in mind. And when they don’t, the cost isn’t theoretical. It shows up in misaligned placements, wasted attention, reputation damages, and media dollars that work harder for your competitors.
Not all AI sees the same landscape.
The advertising ecosystem has a visibility problem. Most tools operate within their own perimeter: their own campaigns, their own tagging system, their own slices of data.
But behavior doesn’t respect boundaries. Users don’t scroll differently because the campaign brief said they would. They don’t engage because a placement looked good in a report.
Understanding how media environments actually perform means training systems on real-world dynamics – scroll patterns, layout friction, ad density, sentiment drift, content fatigue.
It also means scale – not in impressions, but in observed variation. Thousands of layouts. Millions of sessions. Billions of interactions.
Context-rich data. Behaviorally grounded. Competitive by nature.
This is the new cost of entry for decision-making AI.
If agentic AI is going to make choices in your name – about where your ads show, who sees them, and how they land – it should meet a minimum standard:
– It should know what the user is doing.
– It should know how the page behaves.
– It should know what else is on the screen, and why that matters.
– It should learn across environments.
That’s what should be expected from any system that claims to operate with intelligence data.
For brands, the question isn’t “Should we use AI?”
You already are. The question is:
What kind of AI are you trusting to make decisions on your behalf?
Is it trained on assumptions or behavior?
Does it filter based on appearance or learn from experience?
Is it making decisions in isolation or from competitive awareness across the ecosystem?
Confidence in AI will keep growing. But so will the divide between systems that make assumptions at scale, and those that see what’s really happening.
Only one of those paths leads to real advantage.
And the brands who understand that will move both faster and smarter.
Latest Articles
-
Your Media Plan is Still Yours – Until it Isn’t
Platforms like X aren’t building AI to support advertisers - they’re building it to control the process. Systems like Grok optimize for what’s measurable inside their ecosystem, not for a brand’s broader goals. Automation without visibility isn’t strategy, it’s dependency and that shift is one every advertiser should be paying attention to.
View Now -
The Web’s Closing Doors – And AI’s Feeling the Lock
Recently, Cloudflare took the unusual step of publicly calling out Perplexity AI for scraping publisher content without permission. And they didn’t hold back - accusing the company of disguising its crawler, bypassing robots.txt rules, and consuming massive bandwidth without attribution or consent. Cue the familiar debate: who owns online content? Should AI assistants be allowed to “learn” from everything on the web? Is scraping theft - or fair use?
View Now -
We Now Know What Ad Was Served. We Still Don’t Know if it Mattered.
The IAB’s new ACIF standard brings much-needed structure to creative asset tracking in programmatic advertising. It’s a meaningful step — but knowing what was served isn’t the same as knowing if it mattered. This piece explores what ACIF fixes, what it doesn’t, and why understanding the experience of an ad is key to measuring true impact.
View Now