cupure logo
trumpepsteingazadeathtradedealwartrumpsisraelozzy

OpenAI's data center ambitions collide with reality

Data: Axios reporting; Chart: Axios VisualsOpenAI is finding that financing and building the massive data centers it needs to meet its ambitions is easier said than done. Why it matters: There's no limit on AI builders' hunger for computing capacity.Many advances in generative AI have come from dramatically boosting the computing power applies to model training, while new approaches and increased adoption mean that running those models demands more and more data centers, too. Driving the news: The Wall Street Journal reported Monday that OpenAI has run into conflicts and stumbles in its relationship with SoftBank, a key partner in its effort to dramatically scale up its access to computing capacity.The companies pledged in January that they would immediately invest $100 billion.But no specific deals have yet been inked and near-term ambitions have been scaled back, with this year likely to yield only work on a single small U.S. data center, likely in Ohio, per the WSJ.The Journal said other differences have cropped up in the six months since Sam Altman, SoftBank CEO Masayoshi Son and Oracle Chairman Larry Ellison announced their plans with President Trump in the Roosevelt Room of the White House.Yes, but: OpenAI announced Tuesday an expansion of its work with Oracle that will see the pair build a further 4.5 gigawatts of capacity on top of an 800-megawatt project already under development in Abilene, Texas."With this deal, we expect we will exceed the $500 billion commitment we announced at the White House 6 months ago; we are now working to expand that with Oracle, SoftBank, Microsoft, and others," Sam Altman told OpenAI employees Monday in a memo seen by Axios. What they're saying: Altman acknowledged in the employee memo that the company's thirst for computing is "starting to strain the supply chain" and will "require some real creativity."Zoom in: OpenAI, which once relied solely on partner and investor Microsoft, has been partnering with an array of different companies and governments to get more processing power.Its OpenAI for Countries project, announced in May, aims to bring various nations into the mix as partners. It signed its first such deal in May with the United Arab Emirates. As part of OpenAI for Countries, partners will also agree to invest in U.S. projects — helping the company fund yet more data centers here.OpenAI's chief strategy officer Jason Kwon traveled across Asia for three weeks meeting with governments and potential investors in India, Singapore, Korea and other countries, though no additional deals have yet been announced. OpenAI is even turning to rival Google for compute capacity. Altman reassured employees in his memo that additional capacity is coming sooner rather than later. He noted that the company expects to have the equivalent of 2 million Nvidia A100 processors available for use by the end of August and then roughly double that capacity by the end of the year.Between the lines: It's not easy to keep tabs on all the different data center projects OpenAI has announced in recent months.The company has used the Stargate name to describe a number of next-generation data center efforts, but various projects under that banner involve different partners.For example, its Middle East project is dubbed Stargate UAE and involves Oracle, Nvidia, Cisco, SoftBank and G42, a Middle East-based AI startup backed by Microsoft and others.The new project announced Tuesday is also dubbed Stargate, though SoftBank (which actually registered the Stargate trademark) isn't a part of the effort.The big picture: OpenAI isn't the only company with big needs. Global data center capacity is seen growing by 23% per year through 2030, according to a Morgan Stanley report.Much of that growth is being spurred by a handful of big players, including Meta, Google, Microsoft and xAI.Elon Musk said Tuesday that his xAI company has 230,000 GPUs, including 30,000 of Nvidia's GB200, up and running for training Grok in its first Colossus supercluster. Musk said the first batch of a planned 550,000 high-end Nvidia chips, also for training models, will come online in a few weeks. xAI relies on cloud providers to handle running Grok.Anthropic, meanwhile, is acknowledging that meeting its data center needs will require the company to show more flexibility in choosing its partners, including taking money from Middle East governments. "Unfortunately, I think 'No bad person should ever benefit from our success' is a pretty difficult principle to run a business on," Amodei said in a memo to staff, per Wired.The bottom line: All the companies chasing AI with superhuman capabilities say they will need ever more money, more electricity and more processing power."We must continue to get more ambitious about compute; we want to bring lots of AI to lots of people," Altman said in his memo. "It's very clear to me how powerful it's going to be to have AI working for you all of the time and helping you accomplish your goals in all sorts of ways." "One instance of GPT-5 running 24/7 will be powerful," he wrote, "but personally I think 100 instances sounds better."

Comments

World news