A group of prominent mathematicians has released ten unpublished research-level questions to test AI systems' mathematical reasoning capabilities, with answers encrypted to prevent premature solutions.
A consortium of ten prominent mathematicians has released a set of ten unpublished research-level mathematics questions to evaluate the capabilities of current AI systems in solving complex mathematical problems. The questions, which have arisen naturally in the authors' research processes, are being shared publicly for the first time, though their answers remain encrypted for a short period.
The Challenge Setup
The mathematicians—Mohammed Abouzaid, Andrew J. Blumberg, Martin Hairer, Joe Kileel, Tamara G. Kolda, Paul D. Nelson, Daniel Spielman, Nikhil Srivastava, Rachel Ward, Shmuel Weinberger, and Lauren Williams—aim to assess whether current AI systems can correctly answer questions that require genuine mathematical insight rather than pattern matching or memorization.
According to the paper, the questions span multiple mathematical disciplines including algebraic geometry, combinatorics, geometric topology, and ring theory. Each question represents a genuine research problem that the authors have encountered in their work but have not previously published.
Why This Matters
This initiative addresses a growing concern in the mathematical community about the actual capabilities of AI systems when confronted with novel, research-level mathematics. While large language models have demonstrated impressive performance on standardized math tests and competition problems, their ability to contribute to original mathematical research remains uncertain.
The encrypted answer approach is particularly clever—it prevents AI systems from simply searching for solutions online while still allowing the mathematical community to verify whether any AI-generated solutions are correct once the encryption is lifted.
The Broader Context
This challenge comes at a time when AI companies are making increasingly bold claims about their systems' mathematical abilities. Recent models have shown proficiency in solving competition mathematics problems, but research mathematics presents a different challenge—it requires not just computational skill but creative insight and the ability to connect disparate mathematical concepts.
The involvement of Fields Medalists (Hairer) and other highly accomplished mathematicians lends significant credibility to this effort. It represents a rare instance of the mathematical community directly engaging with AI capabilities in a structured, empirical way.
What to Watch For
When the answers are eventually revealed, the mathematical community will be able to assess whether any AI systems have genuinely solved these problems or merely provided plausible-sounding but incorrect responses. This could provide valuable data about the current limitations of AI in mathematical reasoning.
The questions themselves may also prove valuable to the broader mathematical community, as they represent genuine research problems from active areas of mathematical investigation.
Access and Participation
The full set of questions is available through the arXiv preprint server. While the paper doesn't specify how AI systems or their developers can formally participate in the challenge, the mathematical community will likely be watching closely for any claims of successful solutions.

This initiative represents an important step toward empirically understanding the boundaries of AI capabilities in mathematics—a field where correctness is absolute and creativity is paramount.

Comments
Please log in or register to join the discussion