Staying Warm During AI Winter, Part 1: Introduction

Posted by ttkciar@reddit | LocalLLaMA | View on Reddit | 26 comments

The field of AI has always followed boom/bust cycles.

During "AI Summers", advances come quickly and enthusiasm runs high, but commercial interests hype up AI technologies and overpromise on their future capabilities. When those promises fail to materialize, enthusiasm turns to disillusionment, dismay and rejection, and "AI Winter" sets in.

AI Winters do not mark the end of progress in the field, nor even pauses. All manner of technologies developed during past AI Summers are still with us, subject to constant improvement, and even commercial success, but they are not marketed as "AI". Rather, they are called other things -- compilers, databases, search engines, algebraic solvers, provers, and robotics were all once considered "AI" and had their own Summers, just as LLM technology is having its own.

What happens during AI Winters is that grants and venture capital for investing in AI dries up, most (but not all) academics switch to other fields where they can get grants, and commercial vendors relabel their "AI" products as other things -- "business solutions", "analytics", etc. If the profits from selling those products do not cover the costs of maintaining them, those products get shelved. AI startups which cannot effectively monetize their products are acquired by larger companies, or simply shut their doors.

Today's AI Summer shows every sign of perpetuating this pattern. LLM technology is wonderful and useful, but not so wonderful and useful that commercial interests cannot overpromise on its future, which is exactly what LLM service vendors are doing.

If overpromising causes disillusionment, and disillusionment causes AI Winter, then another AI Winter seems inevitable.

So, what does that mean for all of us in the local LLaMa community?

At first glance it would seem that local LLaMa enthusiasts should be in a pretty good position to ride out another Winter. After all, a model downloaded to one's computer has no expiration date, and all of the software we need to make inference happen runs on our own hardware, right? So why should we care?

Maybe we won't, at least for the first year or two, but eventually we will run into problems:

These are all problems which can be solved, but they will be easier to solve, and more satisfactorily, before AI Winter falls, while we still have HF, while Claude and GPT4 are still cheap, while our software is still maintained, and while there are still many eyes reading posts in r/LocalLLaMa.

I was too young to remember the first AI Winter, but was active in the field during the second, and it left an impression on me. Because of that, my approach to LLM tech has been strongly influenced by expectations of another AI Winter. My best guess is that we might see the next AI Winter some time between 2026 and 2029, so we have some time to figure things out.

I'd like to start a series of "Staying Warm During AI Winter" conversations, each focusing on a different problem, so we can talk about solutions and keep track of who is doing what.

This post is just an introduction to the theme, so let's talk about it in general before diving into specifics.