The Agentic Delusion: How AI Research Agents Gave Us More Content, Not Better Cognition

8. March 2025
HCISystem DesignAI InterfacesUser Experience

Most “autonomous research agents” promise to lighten your cognitive load, but usually just hand you more to read. This piece explores why today’s agent tools often increase information overload instead of helping you make sense of it — and why the real challenge is building systems that support actual thinking, not just prettier outputs.

In this article, we will poke into some of the problems with most “agents” today: polished output, static structure. Reports you can read, but not move through — no branching, no reconfiguration, no way to follow a thread when you don’t yet know what you’re looking for.

Failed to load mediasrc: https://kqdcjvdzirlg4kan.public.blob.vercel-storage.com/content/concepts/2025-ephemeral-interfaces/part-2/published/website/images/cover.png

Output up, sense-making down

If you opened Twitter, Substack, Hacker News, Product Hunt or even slow adopter platform like LinkedIn (or whatever your personal choice of bubble pushing “new developments” in technology might be) anytime in the last few months, you’ve probably seen it: a gold rush of “autonomous research agent” startups.

Everywhere you look, someone is pitching their newest Agent startup — services that promise to do the heavy cognitive lifting for you: research, synthesize, summarize, plan, execute. All very Deep™, very Research™.

At first, it sounds like magic. You give the agent a complex task — “compare climate risk models” or “summarize the latest AI safety debates” — it goes to work for a bit… and a few minutes later it hands you back a polished report. Mission accomplished. Work compressed into minutes. Email to your higher-up ready to be sent. Future unlocked.

Except… not really.

Because after the initial dopamine rush of seeing the agent “at work” comes a quieter, slower and more bitter realization: These agents aren’t actually solving the problem of cognitive load. They’re just shifting resposibility downstream — by quietly handing the actual burden of knowledge work back to the user.

You asked for a synthesis. You got a firehose — lightly formatted into paragraphs and headings, dumped into your lap like it’s still 2003.

Instead of a revolution in cognition, what we got was a revolution in output — and the gap between those two things is starting to show.

Text walls

To understand the mess, it helps to be precise about what most “AI agents” today actually do.

When you launch one — whether it’s an agent in a startup demo, a custom GPT you spun up, or an “autonomous researcher” built into a platform — the pattern is basically this:

Step
1Take your query (“analyze X,” “compare Y,” “summarize Z”).
2(Maybe) have some logic that “plans” the agents run and selects tools to call.
3Spawn a dozen subprocesses.
4Scrape whatever data is accessible (web, docs, APIs).
5(Maybe) rank those “sources” in accordance with the initial user task.
6Summarize the data.
7Synthesize the results into a report, but please make it real loooong.
8Hand you a markdown rendering, a PDF, or a chat thread stuffed with bullet points.

The surface looks polished — headings, subheadings, some tables, maybe some “key takeaways” for good measure. But underneath, it’s all the same: linear, brittle, passive.

  • No serious cross-validation of sources.
  • No semantic compression of redundant ideas.
  • No integration and resolution of opposing viewpoints.
  • No meaningful organization beyond “chronological order” or “whatever the agent scraped first.”
  • No real sense-making — just aggregation.

You’re left holding a big static artifact that is intentionally inflated because it looks like you got more for the few cents that you potentially paid (or the minute that you waited): 20 pages of semi-structured findings. It for sure looks like work was done. But now the real work — stitching together coherence, spotting contradictions, building synthesis — falls squarely back onto you.

The agent didn’t really research. It aggregated over the first handful of “sources” it found. It compressed search costs, but it inflated interpretation costs — and disguised that inflation behind a nicely formatted wall of text.

You don’t have an agent helping you think. You have an agent creating cognitive debt faster than you can pay it off.

And because every startup demo “looks clean,” and every founder is quick to call their markdown output an “insight engine,” we pretend the bottleneck has been solved — when really, it’s been pushed deeper into the stack, harder to see and even harder to dig out of.

Overload becomes default

Meanwhile, the reports pile up: Half-read summaries. “Executive briefs” that demand your attention and you keep telling yourself you’ll get to. Bullet-pointed data dumps that feel too dense to even skim.

Once you start interacting with agents at any serious scale, a strange pattern emerges. You don’t feel like you’re getting more done. You feel like you’re drowning.

Every task you delegate spawns more sprawling artifacts: 30 pages of climate policy summaries. 10,000 words on comparative database performance. 400 bullet points on “emerging trends in user interface design.”

At first, you think: “Wow, look how thorough this is.” By the fifth report, you think: “Maybe I should build an agent to summarize all my agent summaries.” (And people are doing exactly that — a grim recursion: agents summarizing agent output summarizing agent output.)

But the problem isn’t just volume. It’s structurelessness. These artifacts are built to look finished — paragraphs, citations, clean headers — but they don’t help you move.

  • No prioritization of what matters most.
  • No signaling of ambiguity or confidence.
  • No pathways through the material depending on your goals.
  • No sense of “where you might want to zoom in, zoom out, or branch sideways.”

Just a giant wall of “knowledge” mulch.

This is the real cognitive bottleneck: not just the amount of information, but the absence of structure that could help you navigate, prioritize, and make sense of it. Volume without structure doesn’t just create backlog — it creates a kind of cognitive quicksand, where every new report adds to the weight but offers no handholds for movement or understanding.

This drags us back to an earlier era — one that’s also aesthetically echoed in the current zeitgeist: it feels like the late 90s/early 2000s internet: static pages, hyperlinks everywhere. No semantic architecture, no flow, no graceful degradation from overview to detail. Similar spirit, similar short-sightedness (hi, Ted). Only this time it’s on steroids: auto-generated, endless, and relentlessly polite. I get the aesthetic nostalgia, but … please.

Each new agent workflow promises to compress labor. Instead, it explodes cognitive backlog — offloading the real interpretive work onto human shoulders, but shrouding it behind a sheen of “done-ness.”

This isn’t just inefficient. It’s regressive. We didn’t solve cognitive overload. We industrialized it.


Markdown to multimedia (same problem)

If today’s agent products are already clogging our cognitive inboxes with walls of markdown, the next phase is easy to predict.

You can almost hear the demo day pitches already:

“We don’t just generate static reports anymore! We generate multimodal outputs! Interactive dashboards! Summarizing videos! Dynamic charts! Data-driven storyboards!”

The “innovation,” if you want to call it that, is simple: Add an x to the .md. Slap some graphs into the text. Auto-generate a video. Maybe synthesize a voice-over if you’re feeling fancy.

Voilà: “Next-Gen Agentic Experiences™.”

It’ll feel fresher, for a moment. The endless walls of text will be broken up with color, motion, different media types. Founders will get to update their landing pages with phrases like “AI-native multimedia intelligence” and “semantic visual storytelling.”

But the structure will remain exactly the same. Big blocks of information. Delivered statically. Designed for consumption, not for movement.

Just shinier bricks on the same crumbling foundation.

(But it has visuals! — Yeah, cool. So did 14th-century illuminated manuscripts. — But they were gorgeous. — They didn’t make it easier to think through trade logistics across medieval Europe.)

And that’s the deeper issue: The hard part was never “make it prettier.” The hard part was always make it explorable — make it navigable when you don’t yet know what you’re looking for.

A dashboard doesn’t solve that. A pie chart doesn’t solve that. An AI-narrated TikTok summarizing a 40-page “deep dive” doesn’t solve that. They just redecorate the bottleneck: moving from unreadable text walls to unclickable charts.

More production isn’t the answer. Better presentation isn’t the answer either.

Because the real problem isn’t just what media the system generates — it’s that sense-making demands a different kind of movement altogether: dynamic reconfiguration, branching exploration, evolving scaffolds that match the user’s unfolding understanding. Passive outputs, no matter how pretty or multimodal, can’t solve the challenge of dynamic cognition.

The real bottleneck is cognitive navigation — the system’s ability to help users move through alternatives, ambiguity, association, contradiction, and evolving understanding.

Without that, every “innovation” in agent outputs is just a slight-of-hand: nice formatting on the same unsolved homework.

Linear vs non-linear presentation

The reason these “enhanced” agent outputs still feel so dead — even when they’re sprinkled with charts and AI-generated voice-overs — is simple: They’re still built for reading, not for thinking.

And there’s a fundamental difference.

Reading is a linear act. You move forward through pre-laid paths: a chapter, a report, a sequence. It’s a passive structure. It works towards a known destination.

Thinking is nonlinear, contextual, and emergent.

You don’t just absorb what’s handed to you.
You bounce between concepts.
You dig into contradictions.
You collapse, expand, recombine.

You need surfaces that yield, adapt, and reorganize — not just more output, but output that can be moved through, restructured, and explored as your understanding evolves.

But most agent-generated outputs today are still static slabs. No matter how you dress them up, they lock you into the system’s frame — or worse, into the agent’s hallucinated structure.

You can’t tug on a thread and watch a conceptual network unfold. You can’t zoom into a contradiction and watch new branches appear. You can’t reorganize the structure itself based on how your understanding is evolving.

You can only scroll — or give up. It’s the 1997 web all over again. Static pages. A few hyperlinks. Maybe a Java applet if someone got ambitious.

The cultural artifacts of “autonomous research agents” today feel eerily similar to the first static websites: impressive at first glance, but suffocating when you try to actually move inside them.

Meanwhile, the real transformation of the web didn’t happen until it became an environment: searchable, dynamic, reconfigurable, navigable across scales and modalities.

In 2025, most AI agent startups are still stuck mentally in Geocities.

We’re cranking out content faster than ever — but we’re moving slower through it, because the structure isn’t keeping up with the scale.

Without true navigation, without surfaces that can reorganize themselves based on intent, AI agents are not solving information overload. They’re manufacturing it, faster.

More text won’t help. Prettier charts won’t help. Summarized videos won’t help.

Only movement will. And movement needs structure. Not fixed scaffolds, but structures that can grow on a shared substrate, bend, and adapt — like thought itself.

Defaults harden

If you zoom out for a second, it’s not surprising that most AI agents today simply hand you a glorified markdown file and call it a day. It’s not just a technical bottleneck — it’s cultural inertia paired with risk aversion and companies lazily milking the hype.

For centuries, the dominant way we externalized complex thought was static text. Books, essays, reports, PDFs, blogs. We spent generations refining techniques for linearizing messy, interconnected knowledge into clean, bounded artifacts. We built deep social rituals around them: publishing, citing, summarizing, concluding — rituals that create a comforting sense of closure, even when the underlying reality is sprawling and ambiguous.

AI agents, despite all their supposed novelty, haven’t escaped this gravity well. They still operate under the quiet assumption that the endpoint of inquiry is a consumable artifact — a static report, a markdown file, something passive that the user reads from start to finish.

And honestly, from a strategic perspective, you can also see why. Static artifacts are cheap. They’re cognitively low-risk. A big markdown file doesn’t crash, doesn’t misrender, doesn’t behave unpredictably halfway through your exploration. It doesn’t demand live orchestration of evolving knowledge. It just sits there — dead, stable, inert.

Low risk for builders. Low confusion for users. Easy to demo. Easy to fund.

But the tragedy runs deeper: the hype machine is rewarding the wrong things — and punishing the right ones.

If you can spin up a 20-page “deep research” report in five minutes, and your early users clap politely and post screenshots, why would you bother tackling the invisible, harder problems? Why sweat over context-preserving navigation, dynamic knowledge orchestration, or scaffolding human sense-making when a slightly fancier markdown export wins you a Series A?

The early “autonomous agents” feel magical not because they deliver structured understanding, but because the bar for what impresses us has quietly sunk to the floor. It “feels good” when a bot scrapes ten papers and stitches them into a report (just as it once felt magical when AltaVista let you search the web at all). It feels “efficient” to get a long ass document in minutes — even if that document leaves you just as overwhelmed and lost as before.

The danger isn’t just that we’re building systems that fail to meet the real needs of human cognition. It’s that we’re getting used to them. Celebrating them Institutionalizing them.

We are normalizing the idea that the solution to overwhelming complexity is… more summaries, more static exports, more pamphlets disguised as progress. We are congratulating ourselves for inventing airplanes — and then refusing to lift off the runway because, oh no, might crash.

And the incentives are clear: When even the strongest players in the space — OpenAI, Anthropic, Perplexity, DeepSeek — are all racing to produce slightly different flavors of “AI search that outputs markdown,” you know the ecosystem is misaligned. (Hint: calling it “semantic retrieval” doesn’t fix much if the final output still looks like a glorified Word doc. I’m just waiting for an animated Mr. Clippy to pop up and comment on its findings in natural voice…)

Without realizing it, we are slowly building a generation of tools optimized not for helping us move through complexity, but for disguising complexity behind prettier formatting.

And the longer this cycle runs, the deeper the assumptions harden — until what once looked like a crude first step quietly cements itself into the default architecture of thought.

Lock-in risk

The real danger isn’t just that today’s agents are simplistic. It’s that their limitations, if left unchallenged, will quietly solidify into the permanent mental models for how we interact with knowledge systems.

Today’s static reports and deterministic workflows feel like “early experiments.” But tomorrow, they could become invisible assumptions — baked into architectures, protocols, interfaces, funding models.

What starts as “just a first step” — a markdown report here, a static dashboard there — risks becoming a structural ceiling.

  • Path Dependency: Early interaction patterns harden. Future systems are expected to output reports, not build living cognitive environments. Funding, design tooling, and roadmaps all bend around the same dead assumptions.
  • Architectural Fossilization: We lock ourselves into workflows optimized for static text dumps, not dynamic thought spaces. Infrastructure (databases, APIs, compliance models) gets designed around exporting “completed artifacts,” not orchestrating live, evolving states.
  • Cognitive Drift: Users internalize bad habits — scanning walls of text, passively consuming surface findings, losing agency over the act of exploration itself.
  • Institutional Lock-in: Enterprises codify these defaults even deeper. “Compliance requires a final report.” “Risk management prefers frozen artifacts.” Suddenly the entire cognitive ecosystem orbits around auditability, not adaptability.

If this trajectory continues, we’ll end up with AI systems beautifully optimized for artifact production — but catastrophically bad at cognition augmentation. We’ll build faster, prettier, denser reports…and still leave users drowning and disconnected.

And here’s the most dangerous part: It will no longer feel wrong. It will feel normal. Natural. Inevitable.

The longer these defaults persist, the more invisible they become — until challenging them feels not just impractical, but irrational. Until the assumptions have hardened so completely that unlearning them becomes harder than building better systems would have been in the first place.

We won’t just have to build new interfaces. We’ll have to dismantle entire conceptual ecosystems: how tools are evaluated, how knowledge is organized, how cognition is scaffolded.

And the window to do that while the clay is still wet — while it’s still flexible — is already narrowing.

Not because today’s agents are failures. But because they are beginnings — and beginnings shape everything that follows.

Beyond today’s agents

This isn’t (just) a rant against agents. It’s a critique of the narrow, chat-bubbles-lead-to-content-dump paradigm that’s become the default — and a call to recognize a much deeper possibility.

Today’s agents are better thought of as automated scribes — impressive at transcribing, summarizing, reassembling knowledge. But they aren’t thinking partners. They don’t shape environments that adapt as we explore. But they could.

If we want agents to actually extend our ability to navigate complexity — not just manufacture more of it — we have to move beyond the idea that the final artifact is the goal.

The real potential lies somewhere else entirely.

Because here’s the thing: the primitives are already here. Every day, we’re glimpsing what could be possible — in ways that the agent hype machine rarely acknowledges.

I’ve gotten plenty of value out of my interactions with LLMs. But almost never when I asked them to “write a report” or “generate a finished piece.” But always when I used them differently:

  • To open new angles on a question.
  • To bounce ideas across perspectives.
  • To surface contradictions I hadn’t yet noticed.
  • To keep moving, twisting, and refining a line of thought — long past the point where a human conversation partner might’ve gotten tired.

The real power of the large language models we converse with isn’t their ability to produce polished artifacts. It’s their relentlessness. The fact that at 4AM, after twenty iterations, they’re still there — still offering reframings, still suggesting new paths, still helping me interrogate my own assumptions.

The real worth is in the interactions themselves: in the turns our conversations take. The detours. The recombinations. The slow emergence of clarity through movement, not closure.

And that’s the fundamental shift we need to embrace:

Artifacts shouldn’t be the endpoint of a task.
They should be openings into deeper investigations.

Not static products to admire, but provisional surfaces to move through — reconfigurable, dissolvable, alive to context and intent.

We need agents that don’t just produce more text, more charts, more exports. We need agents that can materialize environments of thought: spaces that adapt to the friction points, tensions, curiosities, and evolving goals of the humans moving inside them.

That won’t happen by polishing markdown exports a little more. It requires a new architecture. A new interaction model. A new respect for the messy, associative, dynamic nature of real thinking.

It’s harder. It’s riskier. It doesn’t demo as cleanly at startup pitch day. It’s more question than answer.

But it’s the difference between building a tool that feels “productive” — and building one that actually amplifies thought.

The next part of this series explores what that shift demands:

Ephemeral Interfaces — temporary, adaptive interaction spaces that materialize around intent, dissolve after use, and preserve the trails of cognition rather than just some shareable artifact.

Because if we’re serious about building systems that extend thought rather than exhaust it, this is where the real frontier begins.