Random Image Display on Page Reload

Google DeepMind’s AI Agent Dreams Up Algorithms Beyond Human Expertise

May 14, 2025 11:00 AM

Google DeepMind’s AI Agent Dreams Up Algorithms Beyond Human Expertise

A new system that combines Gemini’s coding abilities with an evolutionary approach improves datacenter scheduling and chip design, and fine-tunes large language models.

Image may contain Adult Person Body Part Hand and Text
Photo-Illustration: WIRED Staff/Getty Images

A key question in artificial intelligence is how often models go beyond just regurgitating and remixing what they have learned and produce truly novel ideas or insights.

A new project from Google DeepMind shows that with a few clever tweaks these models can at least surpass human expertise designing certain types of algorithms—including ones that are useful for advancing AI itself.

The company’s latest AI project, called AlphaEvolve, combines the coding skills of its Gemini AI model with a method for testing the effectiveness of new algorithms and an evolutionary method for producing new designs.

AlphaEvolve came up with more efficient algorithms for several kinds of computation, including a method for calculations involving matrices that betters an approach called the Strassen algorithm that has been relied upon for 56 years. The new approach improves the computational efficiency by reducing the number of calculations required to produce a result.

DeepMind also used AlphaEvolve to come up with better algorithms for several real-world problems including scheduling tasks inside datacenters, sketching out the design of computer chips, and optimizing the design of the algorithms used to build large language models like Gemini itself.

“These are three critical elements of the modern AI ecosystem,” says Pushmeet Kohli, head of AI for science at DeepMind. “This superhuman coding agent is able to take on certain tasks and go much beyond what is known in terms of solutions for them.”

Matej Balog, one of the research leads on AlphaEvolve, says that it is often difficult to know if a large language model has come up with a truly novel piece of writing or code, but it is possible to show that no person has come up with a better solution to certain problems. “We have shown very precisely that you can discover something that's provably new and provably correct,” Balog says. “You can be really certain that what you have found couldn't have been in the training data.”

Sanjeev Arora, a scientist at Princeton University specializing in algorithm design, says that the advancements made by AlphaEvolve are relatively small and only apply to algorithms that involve searching through a space of potential answers. But he adds, “Search is a pretty general idea applicable to many settings.”

AI-powered coding is starting to change the way developers and companies write software. The latest AI models make it trivial for novices to build simple apps and websites, and some experienced developers are using AI to automate more of their work.

AlphaEvolve demonstrates the potential for AI to come up with completely novel ideas through continual experimentation and evaluation. DeepMind and other AI companies hope that AI agents will gradually learn to exhibit more general ingenuity in many areas, perhaps eventually generating ingenious solutions to a business problem or novel insights when given a particular problem.

Josh Alman, an assistant professor at Columbia University who works on algorithm design, says that AlphaEvolve does appear to be generating novel ideas rather than remixing stuff it’s learned during training. “It has to be doing something new and not just regurgitating,” he says.

The DeepMind researchers found that they could sometimes give an idea for an algorithm as a prompt and produce interesting new results. Alman says this raises the prospect that human scientists could collaborate with a system like AlphaZero. "That seems really exciting to me," he says.

AlphaEvolve is not the only DeepMind program to demonstrate real ingenuity. The company’s famous board-game-playing program AlphaZero was able to devise original moves and strategies through its own form of experimentation. Balog says that the evolutionary approach used by his group could be coupled with the reinforcement learning method employed in AlphaZero—a process that lets a program learn through positive and negative feedback—to create something that explores new ideas in other areas.

Two previous DeepMind projects also used AI to push the boundaries of computer science. AlphaTensor, from 2022, used the reinforcement learning method to produce novel algorithms. Fun Search, from 2024, used an evolutionary method to generate more efficient code for a given problem.

Neil Thompson, a scientist at MIT who studies the way algorithms affect technological progress, says that a key question is not just whether AI algorithms can exhibit original ideas, but how generally this may apply to scientific research and innovation.

“If these capabilities can be used to tackle bigger, less tightly-scoped problems, it has the potential to accelerate innovation—and thus prosperity,” Thompson says.

What do you make of AlphaEvolve? What novel problems would you like to see AI take on? Let me know by emailing hello@wired.com or adding to the comments section below.

Will Knight is a senior writer for WIRED, covering artificial intelligence. He writes the AI Lab newsletter, a weekly dispatch from beyond the cutting edge of AI—sign up here. He was previously a senior editor at MIT Technology Review, where he wrote about fundamental advances in AI and China’s AI … Read more
Senior Writer

Read More

These Startups Are Building Advanced AI Models Without Data Centers

A new crowd-trained way to develop LLMs over the internet could shake up the AI industry with a giant 100 billion-parameter model later this year.
Will Knight

Singapore’s Vision for AI Safety Bridges the US-China Divide

In a rare moment of global consensus, AI researchers from the US, Europe, and Asia came together in Singapore to form a plan for researching AI risks.
Will Knight

The Middle East Has Entered the AI Group Chat

The UAE and Saudi Arabia are investing billions in US AI infrastructure. The deals could help the US in the AI race against China.
Will Knight

Google’s AI Boss Says Gemini's New Abilities Point the Way to AGI

Google’s AI models are learning to reason, wield agency, and build virtual models of the real world. The company’s AI lead, Demis Hassabis, says all this—and more—will be needed for true AGI.
Will Knight

OpenAI and the FDA Are Holding Talks About Using AI In Drug Evaluation

High-ranking OpenAI employees have met with the FDA multiple times in recent weeks to discuss AI and a project called cderGPT.
Will Knight

Behold the Social Security Administration’s AI Training Video

Social Security workers are being asked to use an AI chatbot. An animated video on how to do so failed to mention that the chatbot can’t be trusted with personally identifiable information.
David Gilbert

‘You Can’t Lick a Badger Twice’: Google Failures Highlight a Fundamental AI Flaw

Google’s AI Overviews feature credible-sounding explanations for completely made-up idioms.
Brian Barrett

AI Is Spreading Old Stereotypes to New Languages and Cultures

Margaret Mitchell, an AI ethics researcher at Hugging Face, tells WIRED about a new dataset designed to test AI models for bias in multiple languages.
Reece Rogers

This Startup Has Created AI-Powered Signing Avatars for the Deaf

New technology from British startup Silence Speaks enables an AI-generated sign language avatar to effectively give the deaf and hard of hearing an interpreter in their pocket.
Simon Hill

With AI Mode, Google Search Is About to Get Even Chattier

Google’s new chatbot-style AI Mode search experience, previously an experiment, is launching for US users. Publishers and marketers will have to adjust their search strategies once again.
Reece Rogers

Take a Tour of All the Essential Features in ChatGPT

If you missed WIRED’s live, subscriber-only Q&A focused on the software features of ChatGPT, hosted by Reece Rogers, you can watch the replay here.
Reece Rogers

Deepfakes, Scams, and the Age of Paranoia

As AI-driven fraud becomes increasingly common, more people feel the need to verify every interaction they have online.
Lauren Goode

*****
Credit belongs to : www.wired.com

Check Also

At Bitcoin 2025, Crypto Purists and the MAGA Faithful Collide

Jessica Klein Business Jun 5, 2025 5:00 AM At Bitcoin 2025, Crypto Purists and the …