THE NEW YORK TIMES: AI technology has solved something not even the smartest humans can - an ‘Erdos problem’

Cade Metz
The New York Times
The latest AI technology has solved an equation no other mathematician has been able to do. But some experts say the AI’s solution was not very different from earlier work done by humans.
The latest AI technology has solved an equation no other mathematician has been able to do. But some experts say the AI’s solution was not very different from earlier work done by humans. Credit: DeltaWorks/Pixabay

For decades, elite mathematicians have struggled to solve a collection of thorny problems posed by a 20th-century academic named Paul Erdos.

This month, an artificial intelligence startup called Harmonic jumped into the mix. Harmonic said its AI technology, Aristotle, had solved an “Erdos problem” with help from a collaborator: OpenAI’s latest technology, GPT-5.2 Pro.

For many computer scientists and mathematicians, solving an Erdos problem showed that AI had reached a point where it was capable of doing legitimate academic research. But some experts were quick to point out that the solution generated by AI was not very different from earlier work done by human mathematicians.

Sign up to The Nightly's newsletters.

Get the first look at the digital newspaper, curated daily stories and breaking headlines delivered to your inbox.

Email Us
By continuing you agree to our Terms and Privacy Policy.

“It feels to me like a really clever student who has memorised everything for the test but doesn’t have a deep understanding of the concept,” said Terence Tao, a professor at UCLA, who is regarded by many as the finest mathematician of his generation. “It has so much background knowledge that it can fake actual understanding.”

The debate over what Harmonic’s system accomplished was a reminder of two consistent questions about the head-spinning progress of the tech industry’s AI development: Did the AI system truly do something brilliant? Or did it merely repeat something that had already been created by brilliant humans?

The answers to those questions could provide a better understanding of the ways AI could transform science and other fields. Whether AI is generating new ideas or not — and whether it may one day do better work than human researchers — it is already becoming a powerful tool when placed in the hands of smart and experienced scientists.

These systems can analyse and store far more information than the human brain, and can deliver information that experts have never seen or have long forgotten.

Dr Derya Unutmaz, a professor at the Jackson Laboratory, a biomedical research institution, said the latest AI systems had reached the point where they would suggest a hypothesis or an experiment that he and his colleagues had not previously considered.

“That is not a discovery. It is a proposal. But it lets you narrow down where you should focus,” said Dr Unutmaz, whose research focuses on cancer and chronic diseases. “It allows you to do five experiments rather than 50. That has a profound, accelerating effect.”

The excitement around GPT-5’s math skills began in October when Kevin Weil, vice president of science at OpenAI, said on social media that the startup’s technology had answered several of the mind-bending Erdos problems.

Designed as a way of measuring mathematical ingenuity, the Erdos problems are elaborate conjectures or questions that test the limits of the field. The aim is to prove whether each is right or wrong.

Some problems are enormously difficult to solve, while others are easier. One of the more famous problems asks: If the integer n is greater than or equal to 2, can 4/n be written as the sum of three positive fractions? In other words, is there a solution to 4/n=1/x+1/y+1/z?

That problem is still unsolved. But on social media, Mr Weil boasted that GPT-5 had cracked many others. “GPT-5 just found solutions to 10 (!) previously unsolved Erdos problems, and made progress on 11 others,” Mr Weil wrote. “These have all been open for decades.”

Mathematicians and AI researchers quickly pointed out that the system had identified existing solutions buried in decades of research papers and textbooks. The OpenAI executive deleted his social media post. But even if the initial excitement was overstated, the technology had proved its worth.

“What it was able to do was surprising and useful,” said Thomas Bloom, a mathematician at the University of Manchester who runs a website dedicated to the Erdos problems. “One of the papers it found was written in German. I never would have found it on my own.”

So how has AI made such leaps since ChatGPT was introduced in late 2022? Today’s leading AI systems are what scientists called neural networks, which can identify patterns in text, sounds and images and learn to generate this kind of material on their own, including term papers, computer code, voices and diagrams.

About 18 months ago, companies like OpenAI and Google started to hone their systems using a technique called reinforcement learning. Through this process, an AI system can learn behaviour through extensive trial and error.

The result is that these systems can now “reason” through problems in fields like maths, science and computer programming. A system like GPT-5 does not exactly reason like a human, but it can spend additional time working on a problem. Sometimes, this work extends for hours.

(The New York Times has sued OpenAI and Microsoft, claiming copyright infringement of news content related to AI systems. The two companies have denied the suit’s claims.)

After Mr Weil’s social media post, researchers continued to ask GPT-5 and other AI technologies for solutions to Erdos problems.

Kevin Barreto and Liam Price, two mathematicians in Britain, used GPT-5 this month to crack a problem that was previously unsolved. Then they used Aristotle, the AI system from Harmonic, to verify that the solution was correct. Unlike GPT-5, Aristotle uses a specialised computer programming language to prove whether an answer is right or wrong.

The two mathematicians did play a small role. They nudged OpenAI’s system in a new direction when its proof did not quite do what they wanted it to do. But like other experts, they believe that AI has now shown that it can do academic research.

Though he was impressed, Mr Bloom added he still hadn’t seen evidence that AI could generate ideas that humans couldn’t. “And I would be surprised if that happened anytime soon,” Mr Bloom said.

Nonetheless, scientists say that AI has become a powerful and rapidly improving research tool and that whether it is generating ideas on its own is — for now — a moot point.

When Dr Unutmaz uses AI for his research into chronic diseases, he said, he often feels like he is talking with an experienced colleague. But he acknowledges the machine cannot do its work without a human collaborator. An experienced researcher is still needed to repeatedly prompt the system, explain what it should be looking for and ultimately separate the interesting information from everything else the system produces.

“I am still relevant, maybe even more relevant,” he said. “You have to have a very deep expertise to appreciate what it is doing.”

This article originally appeared in The New York Times.

© 2026 The New York Times Company

Originally published on The New York Times

Comments

Latest Edition

The Nightly cover for 15-01-2026

Latest Edition

Edition Edition 15 January 202615 January 2026

Albanese’s anti-Semitism bill in dire straits as Greens and Coalition reject legislation.