WriteDown.Blog

The Quiet Cost of AI Overdependence

· Gabriel Coronado
The Quiet Cost of AI Overdependence

AI is one of the most powerful tools we’ve ever had as developers. Used well, it’s a force multiplier. We all know that.

But there’s a quieter effect happening—especially with certain junior and early-stage engineers.

They’re not just using AI. They’re defaulting to it.

Stuck on a bug? Ask AI.
Need a function? Ask AI.
Don’t understand something? Ask AI.

On the surface, that looks like productivity. Underneath, it’s something else: the gradual outsourcing of thinking.

And that’s where things start to break down.


The Difference Isn’t AI Usage—It’s How You Think

I saw this play out clearly in one of my side projects.

We had two junior engineers working on the same codebase. Both relied heavily on AI. But their trajectories couldn’t have been more different.

One would ask questions back. They’d challenge ideas, propose alternatives, and show real curiosity about the problem. When they used AI, it was as a collaborator—something to refine their thinking, not replace it. You could see them engaging with the problem.

The other took the opposite approach. They would copy my question into ChatGPT and return with whatever came out. No validation. No context. No effort to check what already existed in the codebase or whether the solution even made sense. Just a direct pass-through.

Same tool. Completely different mindset.

That difference compounds fast.

Because good engineering isn’t about producing code—it’s about judgment.

Can you tell if a solution is actually correct?
Does it solve the problem you really have?
Is it maintainable? Does it fit the existing system?
What are the trade-offs?

These are the questions that define an engineer. And they’re exactly the ones you skip when you rely on AI without thinking.

AI doesn’t remove the need for judgment—it raises the bar for it. You’re now operating with generated solutions that can look convincing while being subtly wrong, misaligned, or incomplete. If you don’t actively evaluate them, you’re not moving faster—you’re just making mistakes more efficiently.


The Kind of Thinking AI Won’t Give You

There’s another layer to this that people don’t talk about enough.

We tend to overestimate what AI can do—and that’s exactly what makes us lazy in the first place.

AI is very good at operating within a frame.

It’s much worse at questioning the frame itself.

That kind of lateral, “what if” thinking—the kind that challenges assumptions at a fundamental level—is still deeply human.

A great example of this comes from Spotify. [source]

At one point, they were obsessed with reducing playback latency to near zero. The obvious constraint? The internet itself. TCP/IP simply doesn’t allow for that level of responsiveness in the way they needed.

Now imagine asking AI: “How do we reduce latency as much as possible?”

You’d get a clean, well-structured answer:

  • Optimize buffering
  • Use CDNs
  • Compress data
  • Improve caching

All reasonable. All incremental. All trapped inside the same box.

But a real engineer asked a different question:

What if the limitation isn’t something we work around… but something we replace?

Instead of accepting TCP/IP as a given, they explored building their own protocol tailored to their needs.

That’s not an optimization.

That’s a reframing of the problem itself.

And that’s the kind of leap AI almost never gives you—because it’s trained on what exists, not on rejecting it.


The Takeaway

The risk isn’t that AI makes you worse.

It’s that it makes it easier to skip the parts that make you better.

Use AI. You should.

But don’t outsource your brain.

Use it to accelerate your thinking, not replace it.
Interrogate its answers.
Adapt them to your context.
And most importantly—don’t lose the habit of asking better questions.

Because in the end, your value isn’t how quickly you can get an answer.

It’s whether you can recognize if that answer is any good—and when to ignore it entirely.