BANANDRE
NO ONE CARES ABOUT CODE

Navigation

HomeCategories

Categories

Artificial Intelligence(201)
Software Architecture(76)
Software Development(65)
Data Engineering(29)
Engineering Management(21)
Product Management(20)
Enterprise Architecture(8)
← Back to all tags

Tagged with

#long-context

2 articles found

500K Context Fine-Tuning on One GPU: The Breakthrough No One’s Talking About Honestly
Fine-tuning
Featured

500K Context Fine-Tuning on One GPU: The Breakthrough No One’s Talking About Honestly

Unsloth’s new algorithms push LLM context windows to 750K tokens on single GPUs, but the real story isn’t the numbers, it’s what happens when you actually try to use them.

#Fine-tuning#gpu-optimization#LLM...
Read More
Linear Attention’s Revenge: How Kimi Delta Attention Smashes the KV Cache Bottleneck
attention-mechanisms

Linear Attention’s Revenge: How Kimi Delta Attention Smashes the KV Cache Bottleneck

Moonshot AI’s hybrid architecture delivers 6x decoding speed with 75% less memory, making 1M-token contexts actually practical.

#attention-mechanisms#efficiency#LLM...
Read More
BANANDRE
NO ONE CARES ABOUT CODE

Connect

2026 BANANDRE
Privacy PolicyTermsImpressum
Built with 🍌