Oct 15, 2025

6 Posts

A man at a computer says AI ordered pizza, while a delivery man outside holds a fruit basket, highlighting a mix-up.
Oct 15, 2025

DeepSeek Cuts Inference Costs, OpenAI Tightens Ties with AMD, Thinking Machines Simplifies Fine-Tuning, Robots Improve Spatial Awareness

The Batch AI News and Insights: Readers responded with both surprise and agreement last week when I wrote that the single biggest predictor of how rapidly a team makes progress building an AI agent lay in their ability to drive a disciplined process for evals...
A man at a computer says AI ordered pizza, while a delivery man outside holds a fruit basket, highlighting a mix-up.
Oct 15, 2025

Improve Agentic Performance with Evals and Error Analysis, Part 1: When AI agentic systems go astray, it’s tempting to shortcut evals and error analysis. But these processes cas lead to much faster progress.

Readers responded with both surprise and agreement last week when I wrote that the single biggest predictor of how rapidly a team makes progress building an AI agent lay in their ability to drive a disciplined process for evals...
Image shows MolmoAct system processing "Put the plate in the dishwasher" through spatial reasoning.
Oct 15, 2025

Better Spatial Perception for Robots: MolmoAct creates spatial maps for robots to plot their actions before executing text directions

Robot control systems that accept only text input struggle to translate words into motions in space. Researchers developed a system that enables robots to plan spatial paths before they execute text instructions.
Close-up of a violin scroll and pegs, symbolizing precision needed in fine-tuning AI models.
Oct 15, 2025

Fine-Tuning Simplified: Thinking Machines’ new Tinker API makes it easier to fine-tune models on many GPUs

The first offering from Thinking Machines Lab, the startup founded by former OpenAI CTO Mira Murati, aims to simplify — and democratize — the process of fine-tuning AI models.
Graphs compare DeepSeek models showing reduced cost per million tokens with V3.2-Exp over V3.1-Terminus.
Oct 15, 2025

DeepSeek Cuts Inference Costs: DeepSeek-V3.2-Exp streamlines processing using a "lightning indexer," boosting efficiency

DeepSeek’s latest large language model can cut inference costs by more than half and processes long contexts dramatically faster relative to its predecessor.
Stacks of AMD GPUs with OpenAI logo above, illustrating their strategic partnership in processing power.
Oct 15, 2025

OpenAI Strengthens Ties With AMD: OpenAI’s latest multibillion-dollar chip deal would give it 6 gigawatts of computing power and up to 10 percent of AMD

OpenAI, strapped for processing power to drive a worldwide constellation of planned data centers, turned to Nvidia’s archrival AMD.

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox