BANANDRE
NO ONE CARES ABOUT CODE

Navigation

HomeCategories

Categories

Artificial Intelligence(370)
Software Development(183)
Software Architecture(166)
Data Engineering(97)
Engineering Management(55)
Enterprise Architecture(31)
Product Management(27)

Tagged with

#efficiency

4 articles found

Step-3.5-Flash: The 196B Parameter Model That Makes Giants Look Wasteful
efficiency
Featured

Step-3.5-Flash: The 196B Parameter Model That Makes Giants Look Wasteful

Stepfun’s sparse MoE model activates only 11B parameters yet outperforms models 3-5x larger on coding and agentic tasks, delivering 100-300 tok/s on consumer hardware and forcing a reckoning with the parameter count arms race.

#efficiency#moe#sparse-activation...
Read More
Tencent’s 2B-Parameter Youtu-LLM Redefines Efficiency by Outperforming Models 4x Its Size
128k-context

Tencent’s 2B-Parameter Youtu-LLM Redefines Efficiency by Outperforming Models 4x Its Size

Tencent’s Youtu-LLM-2B challenges LLM scaling laws with 128K context and superior agentic capabilities despite having only 1.96B parameters.

#128k-context#efficiency#LLM...
Read More
Linear Attention’s Revenge: How Kimi Delta Attention Smashes the KV Cache Bottleneck
attention-mechanisms

Linear Attention’s Revenge: How Kimi Delta Attention Smashes the KV Cache Bottleneck

Moonshot AI’s hybrid architecture delivers 6x decoding speed with 75% less memory, making 1M-token contexts actually practical.

#attention-mechanisms#efficiency#LLM...
Read More
The AI Scaling Lie: How a 7M-Parameter Model Just Embarrassed Giants Like Gemini and DeepSeek
efficiency

The AI Scaling Lie: How a 7M-Parameter Model Just Embarrassed Giants Like Gemini and DeepSeek

Samsung’s Tiny Recursive Model with microscopic 7M parameters beats massive LLMs on reasoning tasks, challenging the ‘bigger is better’ dogma.

#efficiency#machine-learning#reasoning...
Read More
BANANDRE
NO ONE CARES ABOUT CODE

Connect

2026 BANANDRE
Privacy PolicyTermsImpressum
Built with 🍌