Per Stenström on why we never actually replaced the Von Neumann architecture — and whether we ever will

Posted by WeBeBallin@reddit | programming | View on Reddit | 15 comments

Just interviewed Per Stenström — one of the most prominent computer architects to come out of Europe — and asked him about John Backus's 1977 Turing Award lecture – Backus (inventor of Fortran) coined the term "Von Neumann bottleneck":

That was 49 years ago. Every CPU we've built since has the same architecture.

Per's answer is that the bottleneck never went away — we just got extraordinarily good at hiding it. Cache hierarchies, prefetching, out-of-order execution, speculative execution, cache coherence: the entire post-1980s history of CPU innovation is a stack of workarounds that make the bottleneck invisible for typical workloads without actually removing it.

His take on why we haven't replaced the architecture is essentially legacy — the software ecosystem built on Von Neumann is so vast that migrating to anything fundamentally different would cost decades of investment. His sharper point is that Von Neumann isn't "right" in any absolute sense: the architecture has to be in harmony with the underlying technology, and semiconductors happen to support what Von Neumann needs.

The thread I really wanted his read on was whether we'll ever see a genuine shift away from Von Neumann, or whether AI just pulls another generation of workarounds out of us. After 40+ years in the field he's honestly skeptical. He gave phase change memory as a recent cautionary tale: non-volatile, high-density, performance-competitive with DRAM, Intel and Micron poured huge money into it — and it died because of legacy. Even when a clearly viable alternative shows up, the cost of changing everything built around the current architecture tends to win.

The candidates he treats seriously are processing-in-memory (compute units distributed inside the memory itself — though he was honest this might be Von Neumann with a better layout rather than a genuine break) and entirely new substrates like quantum, which are a different paradigm but probably won't replace classical for general-purpose work.

I’d love a take on this from anyone closer to AI accelerator design or new-substrate work.

Link to full conversation here:

https://www.youtube.com/watch?v=NXVTACHB4Es