OpenAI launched GPT-5.2, its latest model series, focused on professional use. The Thinking and Pro tiers offer major gains in complex reasoning, coding, and accuracy. OpenAI reports GPT-5.2 ...
This study introduces MathEval, a comprehensive benchmarking framework designed to systematically evaluate the mathematical reasoning capabilities of large language models (LLMs). Addressing key ...
When it comes to hard problems, computer scientists seem to be stuck. Consider, for example, the notorious problem of finding the shortest round-trip route that passes through every city on a map ...
Researchers at the University of Science and Technology of China have developed a new reinforcement learning (RL) framework that helps train large language models (LLMs) for complex agentic tasks ...
Artificial intelligence for formal mathematical reasoning startup Harmonic AI Inc. announced today that it has raised $120 million in new funding on a $1.45 billion valuation. The funding is intended ...
Ribbit Capital Leads Round at $1.45B Valuation of Math-Based AI Venture; Emerson Collective Joins Existing Backers Including Sequoia & Kleiner Perkins PALO ALTO, Calif.--(BUSINESS WIRE)--Harmonic, the ...
OpenAI and Google LLC today disclosed that their latest reasoning models achieved gold-level performance in a recent coding competition. The ICPC, as the event is called, is the world’s most ...
A few months before the 2025 International Mathematical Olympiad (IMO) in July, a three-person team at OpenAI made a long bet that they could use the competition’s brutally tough problems to train an ...
Last week, when OpenAI launched GPT-5, it told software engineers the model was designed to be a “true coding collaborator” that excels at generating high-quality code and performing agentic, or ...
A new research paper from Apple details a technique that speeds up large language model responses, while preserving output quality. Here are the details. Traditionally, LLMs generate text one token at ...
Details about OpenAI’s upcoming GPT-5 model have leaked. GitHub accidentally published details of the upcoming model and its four variants in a blog, which was later withdrawn. The leak points to ...