SWE-1.6 introduces significant improvements for model developers, including parallel tool calls and reduced reasoning loops, enhancing daily workflows with a benchmark score matching the previous preview. Senior engineers may find the increased speed of 950 tok/s on the fast tier particularly relevant for optimizing their AI systems.
SWE-1.6 finally feels like the model devs actually want to work with.
Same benchmark score as Preview but parallel tool calls, zero reasoning loops, and way less overthinking.
950 tok/s on the fast tier is going to change how we use Windsurf daily
The Gemini API introduces Flex and Priority service tiers, allowing for cost and latency optimizations for production workloads with minimal changes. This is relevant for engineers looking to enhance their infrastructure efficiency without extensive modifications.
Optimizing continues, today Flex and Priority `service_tiers` for the Gemini API. Optimize costs, reliability and latency for production workloads with a single line change.
**Flex Inference:** Pay 50% less for latency-tolerant workloads (no batch file management) =