Destination

2025-07-04

Apple's claims about large reasoning models face fresh scrutiny from a new study

A new Apple study shows that current reasoning models such as Claude 3.7 Thinking or Deepseek-R1 not only fail with complex logic tasks, but paradoxically even think less with increasing difficulty. The models show three levels of performance: for simple tasks, classic language models without a special thinking function are more precise; for medium complexity, reasoning models have advantages; for high complexity, all models break down completely - regardless of the available computing budget. The researchers speak of a fundamental scaling limit of the reasoning approach and do not see any generalizable problem-so [...]</p>
                    <!-- Buttons -->
			        <div class= Discover Copy

Rating

Innovation

Pricing

Technology

Usability

We have discovered similar tools to what you are looking for. Check out our suggestions for similar AI tools.

venturebeat

2025-10-08

Samsung AI researcher's new, open reasoning model TRM outperforms models 10,000X larger — on specific problems

The trend of AI researchers developing new, small open source generative models that outperform far larger, proprietary peers continued this week with yet another staggering advancement.Alexia Jolicoe [...]

Match Score: 111.72

venturebeat

2025-10-02

'Western Qwen': IBM wows with Granite 4 LLM launch and hybrid Mamba/Transformer architecture

IBM today announced the release of Granite 4.0, the newest generation of its homemade family of open source large language models (LLMs) designed to balance high performance with lower memory and cost [...]

Match Score: 92.46

venturebeat

2025-10-27

MiniMax-M2 is the new king of open source LLMs (especially for agentic tool calling)

Watch out, DeepSeek and Qwen! There's a new king of open source large language models (LLMs), especially when it comes to something enterprises are increasingly valuing: agentic tool use — that [...]

Match Score: 86.54

venturebeat

2025-10-09

Nvidia researchers boost LLMs reasoning skills by getting them to 'think' during pre-training

Researchers at Nvidia have developed a new technique that flips the script on how large language models (LLMs) learn to reason. The method, called reinforcement learning pre-training (RLP), integrates [...]

Match Score: 83.84

venturebeat

2025-10-20

New 'Markovian Thinking' technique unlocks a path to million-token AI reasoning

Researchers at Mila have proposed a new technique that makes large language models (LLMs) vastly more efficient when performing complex reasoning. Called Markovian Thinking, the approach allows LLMs t [...]

Match Score: 82.82

venturebeat

2025-10-08

AI21’s Jamba Reasoning 3B Redefines What “Small” Means in LLMs — 250K Context on a Laptop

The latest addition to the small model wave for enterprises comes from AI21 Labs, which is betting that bringing models to devices will free up traffic in data centers. AI21’s Jamba Reasoning 3B, a [...]

Match Score: 78.17

Destination

2025-09-26

Today's best iPad deals include a record-low price on the latest iPad Air M3

Apple's four iPad models each have their value — the mini is super portable, the standard model with the A16 chip is ideal for casual use while the Pros can handle complex tasks better than som [...]

Match Score: 74.09

Destination

2025-06-07

Apple study finds "a fundamental scaling limitation" in reasoning models' thinking abilities

LLMs designed for reasoning, like Claude 3.7 and Deepseek-R1, are supposed to excel at complex problem-solving by simulating thought processes. But a new study by Apple researchers suggests that these [...]

Match Score: 71.68

venturebeat

2025-10-28

IBM's open source Granite 4.0 Nano AI models are small enough to run locally directly in your browser

In an industry where model size is often seen as a proxy for intelligence, IBM is charting a different course — one that values efficiency over enormity, and accessibility over abstraction.The 114-y [...]

Match Score: 71.59