Scaling Laws for Sparse Mixture-of-Experts at Inference
New routing strategies that reshape how sparse MoE models scale when deployed at inference time.
Curated research from top-tier venues and the reporting we actually read.
New routing strategies that reshape how sparse MoE models scale when deployed at inference time.
Combining process reward models with Monte-Carlo search to dramatically improve multi-step reasoning accuracy.
A low-rank replay mechanism that lets foundation models absorb new domains without losing prior knowledge.
A single representation spanning vision, language, and action for long-horizon robot task planning.
The company claims a 40% reduction in inference cost compared to the previous generation; enterprise API preview opens today.
A joint $1B long-term program focused on alignment and interpretability research.
On math and code tasks the new release rivals closed frontier models — the community response has been enthusiastic.