Building Infrastructure Software in the AI Era: Lower Barriers, Restructured Moats
Over the past while, we've been pushing forward RobustMQ's rule engine and Connector, while also heavily using AI in our development workflow. Throughout this process, one question keeps coming back to me: once AI raises execution speed, what becomes the core competitive advantage in infrastructure software?
This isn't abstract thinking — it's something the actual work keeps forcing me to confront. Before, taking an idea from design to production involved a lot of manual effort. Much of the time wasn't spent "thinking through the problem," but "getting things done." Now AI has driven down that cost, and the pace is noticeably faster. That clarity makes one thing even more apparent: infrastructure software hasn't become simpler because of AI — the hard parts have just moved.
The pressure used to be "can we build it?" Now it's more "can we keep getting it right over the long term?"
Before, much of the time building infrastructure software went into API details, low-level debugging, repetitive glue code, and engineering gruntwork. Those things are genuinely faster now. AI can take on a large portion of the execution-layer work, and the speed at which prototypes and features land is visibly higher. That's real — I feel it every day. But precisely because of this acceleration, the cost of misjudgment has gone up.
The Hard Parts Moved, They Didn't Disappear
The deepest lesson I've taken away lately is that a project's real success or failure depends less and less on "how fast you write" and more on "how precisely you define the boundaries." Whether to build a certain capability, how deep to go, how to design error codes, how to unify rule semantics, how to keep complexity under control as the system evolves — get any of these wrong and AI just amplifies the mistake faster.
Why do I feel this way? Because we can now write code quickly, but the truly hard part is keeping the system running reliably over the long term. A wrong decision might not blow up immediately, but it comes back slowly as technical debt, compatibility burden, and operational complexity. The slowness used to be in implementation; now it's shifted to refactoring and governance.
So I've more or less concluded: the core competitive advantage today is product decisions and system-level judgment — not single-point coding speed. Coding is still very important, but it's more like a foundation capability. Without that foundation, you can't judge whether AI's output is correct, and you can't take over complex problems at critical moments. Especially in infrastructure software, where production behavior, edge cases, and regression risk are all hard constraints — if coding ability is weak, a project is nearly impossible to keep maintainable over time.
Why Senior Engineers Have More Leverage
I used to worry that as AI got stronger, senior developers might be weakened. Working through it, I've found the opposite — senior engineers actually have more leverage. Because the value of seniority was never just "writing code fast." It's problem modeling, risk intuition, architecture tradeoffs, and organizational convergence.
What AI produces is candidate implementations, not final answers. Whoever can judge those candidates with high quality and converge them into a stable system holds the initiative. People who've been through the "manual craftsmanship phase" are usually better at this, because they know where things are most likely to go wrong — and where you have to be conservative.
My own energy increasingly goes into three moves: define boundaries first, make tradeoffs second, then drive execution. Execution can be fast, but if the boundaries and tradeoffs are wrong, speed only makes things more dangerous. This rhythm looks slower, but long-term efficiency is actually much higher.
For Junior Engineers, the Question Isn't "Use AI or Not" — It's "How to Use AI"
I've thought about this a lot, because it matters for the future of teams. The smoother AI tools get, the easier it is for junior engineers to skip foundational training and jump straight to "making it run." Short-term delivery looks fast; long-term, it creates a serious capability gap.
My own view: start using AI early, but don't outsource the thinking. Those "dirty, tedious" processes — manually reproducing bugs, reading logs, tracing root causes, understanding protocol details — are still stages that growth must pass through. Skip them, and you'll enjoy it short-term but end up lacking problem models and engineering instincts.
I'd rather solve this with mechanisms than slogans. For example: require writing your own judgment before looking at AI output; require every fix to come with a minimal reproduction and root cause explanation; require code review to articulate "why this change." These moves look slow, but the long-term payoff is huge. Without them, teams develop a false prosperity: fast delivery, but quality and maintainability keep degrading.
Rust's Asymmetric Advantage in This Era
One more thing I keep coming back to: Rust is developing a kind of "asymmetric advantage" in this era. I'm not a language zealot, and I don't want to advocate for any language — but the engineering reality is that Rust's historical weaknesses were mainly learning curve and hiring cost, and AI is significantly eroding those weaknesses.
Meanwhile, Rust's inherent strengths in safety, performance, and controllable concurrency remain. The result: the weaknesses get shallower, the strengths stay, and the long-term moat becomes clearer. This isn't theoretical — it comes from day-to-day development experience. Before, we'd worry "will Rust slow down the team?" That concern hasn't fully gone away, but it's much weaker now.
Especially in infrastructure software, where long-term stability and maintainability are intrinsically more valuable, AI just makes that path faster to travel. In other words, AI hasn't changed Rust's underlying value — it's just lowered the barrier to accessing it.
Can Infrastructure Software Really Be Written with AI? How to Divide the Work
We've discussed this question many times, and it was one of my earliest concerns. My conclusion: infrastructure software can absolutely be written with AI, and should be — but you can't treat it as "fully automated development." Think of it as a "high-throughput executor."
Why? Because infrastructure software isn't done once features run. Its core is long-term stability, evolvability, and maintainability. AI is already very strong at "implementing a piece of code," but the responsibility for "bearing the system's consequences" always falls on us. Who defines the boundaries, makes the tradeoffs, and takes responsibility for production behavior — none of that can be outsourced.
The division of work I've come to prefer: I'm responsible for direction and constraints; AI is responsible for efficient execution. Architecture boundaries, protocol semantics, error code policy, critical data structures — I make those calls. Boilerplate code, repetitive refactoring, test coverage, documentation, initial drafts of refactors — those are well-suited for AI. This accelerates things without giving up control.
More concretely, I usually split tasks into three layers. The first is the "must not go wrong" core layer: state machine semantics, consistency paths, error handling boundaries — this is always human-led. The second is the "can be fast" implementation layer: interface encapsulation, utility functions, routine adaptation — AI can handle most of this. The third is the "must be validated" convergence layer: regression tests, benchmark results, production observability — humans do the final review.
This division isn't as cool as "fully automated," but long-term results are best. Because it fits a reality: AI handles speed, humans handle direction; AI provides candidates, humans make final judgments. Whether infrastructure software can go the distance ultimately depends not on how much code you can generate in a day, but on whether the system can still evolve stably ten months later.
Final Judgment
If I had to summarize in one sentence: in the AI era, infrastructure software benefits — but that benefit only belongs to those who can close the loop on direction, architecture, tradeoffs, and execution. Code isn't disappearing; code is just shifting from "end goal" back to "means." And the truly expensive capability is becoming: making the right decisions, building the right product, and maintaining it over the long term.
Anxiety will come. I feel it too. Especially seeing "the barrier to writing code drop," it's easy to unconsciously question your own moat. But more important than anxiety is asking yourself one question every day: am I doing pure execution, or am I improving the quality of system decisions? As long as the answer keeps shifting toward the latter, competitive strength is rising.
This may be the most real change in infrastructure software in the AI era: barriers to entry are genuinely falling, but moats haven't disappeared — they've just moved from coding itself to judgment and system capability. For me, that's not bad news. It's actually a clearer direction: put energy into doing the right things, making them happen, and maintaining them well over time.
Project: https://github.com/robustmq/robustmq
