- Updated: March 11, 2026
- 8 min read
Nvidia Commits $26 Billion to Open-Weight AI Models, Shaping the Future of Generative AI
Skip to main contentSecurityPoliticsThe Big StoryBusinessScienceCultureReviewsChevronMoreExpandNewslettersPodcastsVideoLivestreamsMerchSearchSearchCommentLoaderSave StorySave this storyCommentLoaderSave StorySave this storyNvidia will spend $26 billion over the next five years to build open source artificial intelligence models, according to a 2025 financial filing. Executives confirmed the news, which has not been previously reported, in interviews with WIRED.The sizable investment could see Nvidia evolve from a chipmaker with an impressive software stack into a bona fide frontier lab capable of competing with OpenAI and DeepSeek. It’s a strategic move that could further entrench Nvidia’s place as the AI world’s leading chip manufacturer, since the models are tuned to the company’s hardware.Open source models are ones where the weights or the parameters that determine a model’s behavior are released publicly—sometimes with the details of its architecture and training. This allows anyone to download and run it on their own machine or the cloud. In Nvidia’s case, the company also reveals the technical innovations involved in building and training its models, making it easier for startups and researchers to modify and build upon the company’s innovations.On Wednesday, Nvidia also released Nemotron 3 Super, its most capable open-weight AI model to date. The new model has 128 billion parameters (a measure of the model’s size and complexity), making it roughly equivalent to the largest version of OpenAI’s GPT-OSS, though the company claims it outperforms GPT-OSS and other models across several benchmarks.Specifically, Nvidia claims Nemotron 3 Super received a score of 37 on the Artificial Intelligence Index, which scores models across 10 different benchmarks. GPT-OSS scored 33—but several Chinese models scored higher. Nvidia says Nemotron 3 Super was secretly tested on PinchBench, a new benchmark that assesses a model’s ability to control OpenClaw, and ranks number one on that test.Nvidia also introduced a number of technical tricks that it used to train Nemotron 3.These include architectural and training techniques that improve the model’s reasoning abilities, long-context handling, and responsiveness to reinforcement learning.“Nvidia is taking open model development much more seriously,” says Bryan Catanzaro, VP of applied deep learning research at Nvidia. “And we are making a lot of progress.”Open FrontierMeta was the first big AI company to release an open model, Llama, in 2023.CEO Mark Zuckerberg recently rebooted the company’s AI efforts, however, and signaled that it might not make future models fully open. OpenAI offers an open-weight model, called GPT-oss, but it is inferior to the company’s best proprietary offerings, not well-suited to modification.The best US models, from OpenAI, Anthropic, and Google, can be accessed only through the cloud or via a chat interface. By contrast, the weights for many top Chinese models, from DeepSeek, Alibaba, Moonshot AI, Z.ai and MiniMax are released openly and for free. As a result, many startups and researchers around the world are currently building on top of Chinese models.“It’s in our interest to help the ecosystem develop,” says Catanzaro, who joined Nvidia in 2011 and helped spearhead the company’s shift from making graphics cards for gaming to making silicon for AI. Nvidia released the first Nemotron model in November 2023. He adds that Nvidia recently finished pretraining a 550-billion-parameter model.(Pretraining involves feeding huge quantities of data into a model spread across vast numbers of specialized chips running in parallel.) Nvidia has since released a range of models specialized for use in areas like robotics, climate modelling, and protein folding.Kari Briski, VP of generative AI software for enterprise, says Nvidia’s future AI models will help the company improve not just its chips but also the super-computer-scale datacenters it builds.“We build it to stretch our systems and test not just the compute but also the storage and networking, and to kind of build out our hardware architecture roadmap,” she says.Releasing models openly may have long-term strategic benefits for Nvidia, too. The company’s chips remain the gold standard for training large AI models, with customers spending billions to acquire the company’s hardware for their datacenters.But the rise of Chinese open models might at some point erode Nvidia’s position if those models were to demonstrate dramatic improvements on rival hardware.In January 2025, DeepSeek released a cutting-edge open model using a more efficient approach that made its training far cheaper. But a variety of other Chinese models from big companies like Alibaba, as well as startups like Moonshot AI, Z.ai, and MiniMax, have also become popular in the West.Alibaba’s model Qwen, which is easy to use and modify and is well maintained, is widely used by researchers and startups.A new DeepSeek model, expected to be released soon, is widely rumored to have been trained exclusively on chips made by the Chinese company Huawei, which is subject to US government sanctions. If true, the release could prompt more startups and researchers to try Huawei’s hardware, particularly in China.In this respect, Nvidia may help shape AI competition between the US and China by providing a US-made alternative to open-weight Chinese models.“We’re an American company, but we work with companies across the world,” Catanzaro says. “It’s in our interest to make the ecosystem diverse and strong everywhere.”Some industry experts have warned that seeing open innovation shift to the other side of the world could be bad for the US in the long run.“I’m a huge Nemotron fan,” says Nathan Lambert, an AI researcher at the Allen Institute for AI (Ai2) who leads the ATOM (American Truly Open Models) Project. Lambert adds that the US government should also fund open models.Andy Konwinski, a computer scientist and entrepreneur who leads the Laude Institute, a nonprofit focused on promoting openness in AI, says Nvidia’s investment is highly significant because of its position at the nexus of AI research.“They sit at the front of so many open and closed AI efforts,” Konwinski says. “This is an unprecedented signal of their belief in openness.”CommentsBack to topTriangleYou Might Also LikeIn your inbox: Our biggest stories, handpicked for you each dayWhat a Google subpoena response looks like—courtesy of the Epstein filesBig Story: The undersea cable that made the global internet possibleEveryone speaks incel nowReplay: Livestream on the hype, reality, and future of EVsWill Knight is a senior writer for WIRED, covering artificial intelligence.He writes the AI Lab newsletter, a weekly dispatch from beyond the cutting edge of AI—sign up here. He was previously a senior editor at MIT Technology Review, where he wrote about fundamental advances in AI and China’s AI . Read MoreSenior WriterXTopicsAI Labartificial intelligenceNVIDIAopen sourceChinaNvidia’s Deal With Meta Signals a New Era in Computing PowerThe days of tech giants buying up discrete chips are over. AI companies now need GPUs, CPUs, and everything in between.Big Tech Says Generative AI Will Save the Planet. It Doesn’t Offer Much ProofA new report finds that of 154 specific claims about how AI will benefit the climate, just a quarter cited academic research. A third included no evidence at all.Code Metal Raises $125 Million to Rewrite the Defense Industry’s Code With AIThe Boston startup uses AI to translate and verify legacy software for defense contractors, arguing modernization can’t come at the cost of new bugs.AI Industry Rivals Are Teaming Up on a Startup AcceleratorOpenAI, Anthropic, Google, and a host of other major tech companies have found common ground in F/ai, a new startup accelerator based out of Paris.Nvidia Is Planning to Launch an Open-Source AI Agent PlatformAhead of its annual developer conference, Nvidia is readying a new approach to software that embraces AI agents similar to OpenClaw.OpenAI Announces Major Expansion of London OfficeThe San Francisco-based AI lab is growing its research team in London. The move puts it in direct competition with Google DeepMind for top research talent in the UK.Inside OpenAI’s Race to Catch Up to Claude CodeWhy is the biggest name in AI late to the AI coding revolution?What AI Models for War Actually Look LikeWhile companies like Anthropic debate limits on military uses of AI, Smack Technologies is training models to plan battlefield operations.ByteDance’s AI Ambitions Are Being Hampered by Compute Restraints and Copyright ConcernsByteDance’s new Seedance 2.0 AI video model seemed unstoppable—until heavy demand strained the company’s compute capacity and copyright complaints began piling up.AI Will Never Be ConsciousIn his new book, A World Appears, Michael Pollan argues that artificial intelligence can do many things—it just can’t be a person.This Defense Company Made AI Agents That Blow Things UpScout AI is using technology borrowed from the AI industry to power lethal weapons—and recently demonstrated its explosive potential.What Is That Mysterious Metallic Device US Chief Design Officer Joe Gebbia Is Using?Gebbia was reportedly spotted at a San Francisco coffee shop using an unidentified pair of earbuds with a circular disc that looks similar to the device seen in a recent OpenAI hoax ad.