Save StorySave this storySave StorySave this story
Ever since DeepSeek burst onto the scene in January, momentum has grown around open source Chinese artificial intelligence models. Some researchers are pushing for an even more open approach to building AI that allows model-making to be distributed across the globe.
Prime Intellect, a startup specializing in decentralized AI, is currently training a frontier large language model, called INTELLECT-3, using a new kind of distributed reinforcement learning for fine-tuning. The model will demonstrate a new way to build competitive open AI models using a range of hardware in different locations in a way that does not rely on big tech companies, says Vincent Weisser, the company’s CEO.
Weisser says that the AI world is currently divided between those who rely on closed US models and those who use open Chinese offerings. The technology Prime Intellect is developing democratizes AI by letting more people build and modify advanced AI for themselves.
Improving AI models is no longer a matter of just ramping up training data and compute. Today’s frontier models use reinforcement learning to improve after the pre-training process is complete. Want your model to excel at math, answer legal questions, or play Sudoku? Have it improve itself by practicing in an environment where you can measure success and failure.
“These reinforcement learning environments are now the bottleneck to really scaling capabilities,” Weisser tells me.
Prime Intellect has created a framework that lets anyone create a reinforcement learning environment customized for a particular task. The company is combining the best environments created by its own team and the community to tune INTELLECT-3.
I tried running an environment for solving Wordle puzzles, created by Prime Intellect researcher, Will Brown, watching as a small model solved Wordle puzzles (it was more methodical than me, to be honest). If I were an AI researcher trying to improve a model, I would spin up a bunch of GPUs and have the model practice over and over while a reinforcement learning algorithm modified its weights, thus turning the model into a Wordle master.
Although reinforcement learning is now incredibly important, it is mostly done behind closed doors by big AI companies. The process normally requires a lot of expertise, putting it out of the reach of most companies and developers. Weisser says that allowing startups to do their own reinforcement learning could produce valuable new software products including agents specialized for all sorts of tasks.
Some experts agree. Andrej Karpathy, the former head of Tesla’s AI team, described Prime Intellect’s reinforcement learning environments as “a great effort [and] idea,” shortly after they were announced. He encouraged open source researchers to take different environments and adapt them to new tasks to improve the skills of advanced models in new ways.
Prime Intellect has already shown that distributed methods—including dividing up calculations and then combining them to create a single, larger model—can challenge conventional ways of building AI. In late 2024 the company announced INTELLECT-1, a 10-billion-parameter model trained with distributed hardware. In March, it unveiled a larger, more capable model, INTELLECT-2, with reasoning capabilities enabled by distributed reinforcement learning.
The AI landscape has shifted dramatically over the past two years. Meta kicked off the open source AI era by releasing the first version of its Llama model in 2023, but the company’s latest offering, announced in April 2025, was a huge disappointment. Meanwhile DeepSeek, a little-known Chinese upstart, shocked the world by unveiling a capable, low-cost reasoning model in January of 2025. Several other Chinese AI models have followed suit. OpenAI responded to DeepSeek’s success this August by launching its first open source model in several years but Chinese models like Alibaba’s Qwen, Kimi from Moonshot, and DeepSeek’s R1 have proved more popular, perhaps because they are easy to modify and adapt.
“It’s almost like the US is out of options when it comes to open frontier models,” Weisser told me. “That's one of the things that we are trying to change.”
What do you think of the Prime Intellect approach? Send an email to ailab@wired.com to let me know.
This is an edition of Will Knight’s AI Lab newsletter. Read previous newsletters here.