Skip to main content
← Back to Blog

Performance

5 articles

A desktop workstation running a local LLM for coding — 80 billion parameters, 3 billion active — representing the shift from cloud AI to local AI coding
AI & Developer Experience

The Local LLM Coding Revolution Just Started — 80B Parameters on Your Desktop, 3B Active, Zero Cloud Bills

A tech journalist just declared he finally found a local LLM he wants to use for real coding work. Qwen3-Coder-Next runs 80 billion parameters on a desktop, activates only 3 billion per token, and plugs directly into Claude Code. The cloud-only era of AI coding is ending. Here is the full technical breakdown, the privacy argument nobody is making, and why this changes the economics of AI-assisted development.

18 min read
Read more →
Node.js logo — V8 pointer compression cuts Node.js memory usage in half with a single Docker image swap
Node.js

Node.js Just Cut Its Memory in Half — One Docker Line, Zero Code Changes, $300K Saved

V8 pointer compression finally comes to Node.js after 6 years. A single Docker image swap drops heap memory by 50%, improves P99 latency by 7%, and can save companies $80K-$300K/year. Cloudflare, Igalia, and Platformatic collaborated to make it happen. Here is the full technical breakdown, real production benchmarks on AWS EKS, and why your CFO needs to see this.

22 min read
Read more →