venturebeat
Open source Mamba 3 arrives to surpass Transformer architecture with nearly 4% improved language modeling, reduced latency

The generative AI era began for most people with the launch of OpenAI's ChatGPT in late 2022, but the underlying technology — the "Transformer" neural network architecture that allows AI models to weigh the importance of different words in a sentence (or pixels in an image) differently and train on information in parallel — dates back to Google's seminal 2017 paper "Attention Is All You Need."Yet while Transformers deliver unparalleled model quality and have underpinned most of the major generative AI models used today, they are computationally gluttonous. They are burdened by quadratic compute and linear memory demands that make large-scale inference an expensive, often prohibitive, endeavor. Hence, the desire by some researchers to improve on them by dev [...]

Rating

Innovation

Pricing

Technology

Usability

We have discovered similar tools to what you are looking for. Check out our suggestions for similar AI tools.

venturebeat
'Western Qwen': IBM wows with Granite 4 LLM launch and hybrid Mamba/Transformer architecture

IBM today announced the release of Granite 4.0, the newest generation of its homemade family of open source large language models (LLMs) designed to balance high performance with lower memory and cost [...]

Match Score: 346.44

venturebeat
Attention ISN'T all you need?! New Qwen3 variant Brumby-14B-Base leverages Power Retention technique

When the transformer architecture was introduced in 2017 in the now seminal Google paper "Attention Is All You Need," it became an instant cornerstone of modern artificial intelligence. Ever [...]

Match Score: 266.66

venturebeat
TII’s Falcon H1R 7B can out-reason models up to 7x its size — and it’s (mostly) open

For the last two years, the prevailing logic in generative AI has been one of brute force: if you want better reasoning, you need a bigger model. While "small" models (under 10 billion param [...]

Match Score: 187.97

venturebeat
IBM's open source Granite 4.0 Nano AI models are small enough to run locally directly in your browser

In an industry where model size is often seen as a proxy for intelligence, IBM is charting a different course — one that values efficiency over enormity, and accessibility over abstraction.The 114-y [...]

Match Score: 139.64

venturebeat
New ‘Test-Time Training’ method lets AI keep learning without exploding inference costs

A new study from researchers at Stanford University and Nvidia proposes a way for AI models to keep learning after deployment — without increasing inference costs. For enterprise agents that have to [...]

Match Score: 109.07

venturebeat
Nvidia debuts Nemotron 3 with hybrid MoE and Mamba-Transformer to drive efficient agentic AI

Nvidia launched the new version of its frontier models, Nemotron 3, by leaning in on a model architecture that the world’s most valuable company said offers more accuracy and reliability for agents. [...]

Match Score: 108.31

venturebeat
The enterprise voice AI split: Why architecture — not model quality — defines your compliance posture

For the past year, enterprise decision-makers have faced a rigid architectural trade-off in voice AI: adopt a "Native" speech-to-speech (S2S) model for speed and emotional fidelity, or stick [...]

Match Score: 106.10

venturebeat
Black Forest Labs launches Flux.2 AI image models to challenge Nano Banana Pro and Midjourney

It's not just Google's Gemini 3, Nano Banana Pro, and Anthropic's Claude Opus 4.5 we have to be thankful for this year around the Thanksgiving holiday here in the U.S.No, today the Germ [...]

Match Score: 101.27

venturebeat
Nvidia's new open weights Nemotron 3 super combines three different architectures to beat gpt-oss and Qwen in throughput

Multi-agent systems, designed to handle long-horizon tasks like software engineering or cybersecurity triaging, can generate up to 15 times the token volume of standard chats — threatening their cos [...]

Match Score: 88.72