

I’ve been researching this a bit… I’ve come to the conclusion that there is no AI bubble. In fact, we’re only just getting started down this road. Unless there’s some massive 100x efficiency breakthrough in training AI and inference, the entire world is going to be building seemingly endless AI data centers (and the normal compute kind, e.g. for stuff like AWS, Google/YouTube, Meta, banks) for at least a decade. Probably a little longer (12-15 years before demand levels out).
Everyone thinks that “AI data center” means ChatGPT, Claude, Gemini, etc but there’s 10,000x more demand for AI than those services. Think: Pharmaceutical companies trying to find proteins, scientists (and big agriculture!) trying to model the weather, and other businesses trying to automate stuff. Not just software; robots and things like conveyor belts.
Another example: Ever use one of those self-checkouts that’s mostly just a camera pointing down, where you place the stuff you’re purchasing? That uses AI too.
Having said that, there is a great big bubble in AI: OpenAI, specifically. That will definitely pop one day. And hopefully, the DRAM bullshit will go along with it.









The reason why you hear this so often is because academia is designed to teach students based on a logical, reasonable curriculum. The curriculum will be mostly well-thought-out and cover all the important topics.
Then you take someone who followed this perfectly reasonable path and you place them in front of the total shitshow that is most businesses. Everything they think they know won’t be applicable because most of the time, logic and reason were not what drove adoption of any given tool or practice.