Reaching AI Limits (and Ending Up on StackOverflow Again)

Contributing to StackOverflow (and the public domain) still matters — just not mainly for humans anymore, but for the AI systems that serve humans.

Reaching AI Limits (and Ending Up on StackOverflow Again)

I use AI daily; it boosts my productivity.

Until it doesn't.

A trivial case: try searching the "cast send blob", and you'll find my self-answered StackOverflow post at the top.

Self-answered post!
Why?

Why is it at the top, and why is it self-answered?

That post exists for a simple reason: AI failed me on a trivial edge case, wasting a few hours ⏳ of engineering time.

ethereum.stackexchange.com is a StackOverflow for the Ethereum-related topics

Once I found a solution, I decided to contribute to public knowledge, thus posting a self-answered solution, allowing the search engine to index it and filling the "knowledge gap".

But again...

Why did the issue appear at the first place?

Fast forwarding: Nobody asks publicly → Nobody answers → AI knows nothing.

Once the article appears publicly, Google indexes it, and it quickly appears at the top. Now, AI picks it up and even partially uses my wording.

Public Knowledge Is Decaying

What happens is that in the current circumstances, when everybody is locked in their AI microcosms, information stops recycling through the public domain.

It's well-known that "StackOverflow is dying" (as for 2026), etc.

Stack Overflow's decline — Eric Holscher

But I see it differently.

As the internet floods with AI-generated code, the "Solved questions" on StackOverflow are transitioning from "Help Desk answers" to "Ground Truth." Human validation is becoming the premium service.

(from the article here: https://gist.github.com/tanaikech/30b1fc76da0da8ff82b91af29c0c9f83)

The incentive to contribute to public knowledge has shifted. Now it is less like helping people directly and more like feeding future models that everyone will use.

AI Wastes Engineering Time at the Edges

I spent over an hour ⏳ asking different LLM models from different vendors how to send a blob via the cast tool. All models answered confidently – and all were wrong. They invented CLI flags, hallucinated parameters, and ignored requests for up-to-date documentation.

The real solution took minutes ⏳: I stopped prompting and read cast --help carefully (and then I posted that self-answer).

The other day, it happened with LogQL (a query language for Grafana Loki). After an hour or two ⏳ of back-and-forth with AI, I gave up, opened the official docs, and composed a proper LogQL query myself.

This happened to me a few times, and it doesn't look random.

LLMs fail similarly at edge cases.

Edge Cases Break LLM Reasoning

LLMs are trained on public texts (particularly). When edge cases aren't documented, models don't say "I don't know". They struggle to find correct solutions and often hallucinate non-working answers overconfidently.

This failure mode is kind of invisible. If you already know the tool or the domain well, you'll notice when the answer smells wrong and go read the docs. If you don't, you might just keep trying random variations of a hallucinated solution, assuming you're the one missing something.

That's not great, especially for less popular tools, smaller projects, or newer ecosystems. It's wasted engineering time, and it adds up. The problem isn't that the models don't know. It's that they don't signal uncertainty. The former might be fixable – models can be taught to say "I don't know". The dependency on public knowledge is a different thing, though. It's not easy to "fix".


It turns out both me and AI end up relying on static non-conversational public knowledge.

In a way, this brings us back to a familiar place: contributing to public knowledge suddenly matters. Maybe not (only) because "open source matters", but because AI depends on it.

It's just that contributing to the public domain matters not only to humans but also to AI serving humans.