Recent Stories

Fresh off the press

Agents & Frameworks

ACL 2026: Dense Communication Topologies in Multi-Agent LLM Systems Accelerate Premature Convergence — and Adding More Agents Makes It Worse

An ACL 2026 Findings paper shows dense communication topologies in [multi-agent LLM systems](/articles/neural-computers-symbolic-stability-failure-contradicts-the-case-for-pure/) accelerate premature convergence, meaning topology matters more than model strength.

8 min read
Read
Agents & Frameworks

'Beyond the Diff' Quantifies Agentic Entropy — Why AI Coding Agents Drift From Intent Across Iteration Steps Even When Each Diff Passes Review

A CHI 2026 paper formalizes agentic entropy as structural drift between agent actions and intent, showing why per-step benchmarks miss cumulative misalignment in long agent.

8 min read
Read
Industry & Business

CATL's 10-to-98%-in-Seven-Minute LFP Cell Pushes the EV Fast-Charge Bottleneck From Battery to Charger Grid

CATL's Shenxing LFP claims 10-to-98% in 6:27, implying ~700–900 kW sustained draw that exceeds CCS1 and Tesla V4 limits and shifts the fast-charging bottleneck from cell.

8 min read
Read
Infrastructure & Runtime

CoCoDiff Exposes the All-to-All Bottleneck That Caps Distributed Diffusion Transformer Inference Well Below Theoretical GPU Count

Ulysses parallelism caps distributed DiT inference scaling on heterogeneous interconnects. CoCoDiff delivers 3.6x average speedups on Aurora via topology-aware scheduling.

8 min read
Read
Agents & Frameworks

Diversity Collapse in Multi-Agent LLM Systems: Structural Coupling Breaks Open-Ended Idea Generation Even When Topologies Are Sparse

An ACL 2026 Findings paper finds multi-agent LLM brainstorming collapses because agents share models, prompts, and context, not because topologies are too dense.

8 min read
Read
Models & Research

DuQuant++ Brings Fine-Grained Rotation to FP4: What Microscaling Quantization Means for Running Larger Models on the Same GPU

DuQuant++ adapts outlier-aware rotation to MXFP4, halving online rotation cost on LLaMA 3 and shifting the FP4 deployment bottleneck from memory to calibration engineering.

8 min read
Read

Stay Ahead of the Curve

Get the latest AI and tech insights delivered to your feed.