Destination

2025-05-06

The Rise of Mixture-of-Experts: How Sparse AI Models Are Shaping the Future of Machine Learning

Mixture-of-Experts (MoE) models are revolutionizing the way we scale AI. By activating only a subset of a model’s components at any given time, MoEs offer a novel approach to managing the trade-off between model size and computational efficiency. Unlike traditional dense models that use all parameters for every input, MoEs achieve enormous parameter counts while […]


The post Discover Copy

Rating

Innovation

Pricing

Technology

Usability

We have discovered similar tools to what you are looking for. Check out our suggestions for similar AI tools.

venturebeat

2025-09-29

DeepSeek's new V3.2-Exp model cuts API pricing in half to less than 3 cents per 1M input tokens

DeepSeek continues to push the frontier of generative AI...in this case, in terms of affordability.The company has unveiled its latest experimental large language model (LLM), DeepSeek-V3.2-Exp, that [...]

Match Score: 82.16

venturebeat

2025-10-27

MiniMax-M2 is the new king of open source LLMs (especially for agentic tool calling)

Watch out, DeepSeek and Qwen! There's a new king of open source large language models (LLMs), especially when it comes to something enterprises are increasingly valuing: agentic tool use — that [...]

Match Score: 77.44

venturebeat

2025-10-02

'Western Qwen': IBM wows with Granite 4 LLM launch and hybrid Mamba/Transformer architecture

IBM today announced the release of Granite 4.0, the newest generation of its homemade family of open source large language models (LLMs) designed to balance high performance with lower memory and cost [...]

Match Score: 68.59

Destination

2025-06-15

Rednote releases its first open-source LLM with a Mixture-of-Experts architecture

Social media company Rednote has released its first open-source large language model. The Mixture-of-Experts (MoE) system, called dots.llm1, is designed to match the performance of competing models at [...]

Match Score: 61.31

venturebeat

2025-10-13

Self-improving language models are becoming reality with MIT's updated SEAL technique

Researchers at the Massachusetts Institute of Technology (MIT) are gaining renewed attention for developing and open sourcing a technique that allows large language models (LLMs) — like those underp [...]

Match Score: 60.62

venturebeat

2025-10-24

Thinking Machines challenges OpenAI's AI scaling strategy: 'First superintelligence will be a superhuman learner'

While the world's leading artificial intelligence companies race to build ever-larger models, betting billions that scale alone will unlock artificial general intelligence, a researcher at one of [...]

Match Score: 58.18

venturebeat

2025-10-29

Vibe coding platform Cursor releases first in-house LLM, Composer, promising 4X speed boost

The vibe coding tool Cursor, from startup Anysphere, has introduced Composer, its first in-house, proprietary coding large language model (LLM) as part of its Cursor 2.0 platform update. Composer is d [...]

Match Score: 49.72

venturebeat

2025-10-28

IBM's open source Granite 4.0 Nano AI models are small enough to run locally directly in your browser

In an industry where model size is often seen as a proxy for intelligence, IBM is charting a different course — one that values efficiency over enormity, and accessibility over abstraction.The 114-y [...]

Match Score: 48.93

venturebeat

2025-10-17

Cisco warns enterprises: Without tapping machine data, your AI strategy is incomplete

Cisco executives make the case that the distinction between product and model companies is disappearing, and that accessing the 55% of enterprise data growth that current AI ignores will separate winn [...]

Match Score: 44.01