Meta has released its first model from the Superintelligence Labs, which may indicate a shift in their AI strategy. Senior engineers should evaluate its capabilities and potential integration into existing systems.
Top stories in AI today:
- Meta Superintelligence Labs ships first model
- HeyGenβs Avatar V solves AIβs identity drift
- Build an automated ad generator with this tool
- Anthropic simplifies the agent-building system
- 4 new AI tools, community workflows, and more
This tweet discusses a novel soft robot design that utilizes heat-responsive materials and embedded electronics for movement without traditional mechanical systems. Senior engineers may find the innovative approach to robotics and materials science relevant for future applications in AI and automation.
A new origami-inspired soft robot uses heat-responsive materials and embedded electronics to move, fold, and reshape itself, without motors, pumps, or bulky mechanical systems.
@Princeton
The tweet discusses Aave's transition plan to shift risk management to decentralized infrastructure, highlighting a significant move in DeFi. Senior engineers should note the implications for on-chain finance and risk management systems.
If you believe global finance belongs onchain, you cannot rely on centralized, off-chain risk silos.
@LlamaRisk
βs transition plan for Aave shifts risk management to neutral, trusted infrastructure.
DeFi will win with
@aave
V4.
The tweet highlights the adoption of Chinese open source AI models by notable companies like Cursor and Cognition, indicating a shift in the AI landscape. Senior engineers should note the implications of this trend on competition and innovation in AI infrastructure.
Silicon Valley is quietly running on Chinese open source AI models.
Here are the receipts:
β Cursor confirmed last month that Composer 2 is built on Moonshot's Kimi K2.5
β Cognition's SWE-1.6 model is likely post-trained on Zhipu's GLM
β Shopify saved $5M a year by
π 9,371 viewsβ€ 48π 5π¬ 13π 230.7% eng
The tweet discusses Gemma 4's use of shared KV cache layers, which allows it to run on a laptop but also highlights a limitation in cache reuse for llama.cpp. This insight into architecture could be relevant for engineers working on efficient AI system designs.
There is a catch nobody is talking about.
Gemma 4 uses shared KV cache layers - the last layers reuse K/V tensors from earlier layers instead of computing their own. That is why it fits on a laptop.
But that same architecture breaks cache reuse in llama.cpp. Every request
π 5,927 viewsβ€ 33π 9π¬ 10π 390.9% eng
ConvApparel is a new dataset aimed at improving LLM-based user simulators by quantifying the 'realism gap.' This could be relevant for engineers focused on enhancing conversational agent training methodologies.
Introducing ConvApparel, a new human-AI conversation dataset, as well as a comprehensive evaluation framework designed to quantify the "realism gap" in LLM-based user simulators and improve the training of robust conversational agents.
Read all about it β
goo.gle/41k5eff
KellyBench tested frontier AI models in a simulated betting market, revealing that all models lost money, with varying degrees of ROI. This highlights the challenges and limitations of current AI models in real-world applications, which is crucial for engineers to consider.
Interesting new benchmark called KellyBench which put frontier models in a simulated Premier League betting market for a full season. Every model lost money.
- Claude Opus 4.6: -11% mean ROI, avoided ruin
- GPT-5.4: -13.6% mean ROI, avoided ruin
- Grok 4.20: -88.2% ROI, went
Fastly's integration of Compute and Semantic Caching optimizes AI agent performance by reducing operational costs at the network edge. This could be relevant for engineers looking to improve the efficiency of deploying AI models in production environments.
$FSLY Fastly optimizes Claude Managed Agents by moving intelligence to the network edge. Integrating Fastly Compute and Semantic Caching significantly lowers the cost of running frontier models / AI agents. Claude Opus 4.6 charges per token for every interaction, for example.