Frontier Research — OpenAI

Graviton

Reading OpenAI's Graviton paper for what it implies about the cost curve of running agents in production — not just training models.

OpenAI Research·6 min read·1 primary source

01Why an operator reads "Graviton"

Most operators skip OpenAI's deeply technical papers — including "Graviton." They shouldn't. The infrastructure papers are the leading indicator for what becomes affordable to deploy in production six to twelve months later.

"Graviton" sits squarely in that category. Its specific contribution is engineering. Its broader implication is about the cost-per-operation curve — and that curve is the single most important variable for whether a given workflow is economical to put a model on.

02The wrong unit of analysis

Most teams budget for AI in cost-per-token. That's the wrong unit. The right unit is cost-per-decision, or cost-per-resolved-ticket, or cost-per-document-shipped — and those numbers move very differently from token cost.

A workflow where the model takes ten attempts to get one decision right is ten times more expensive than the per-token cost suggests. As per-token cost drops, cost-per-decision drops faster, because the affordable retry-and-verify budget grows. That's the curve that actually decides whether automation is economical, and almost no one is graphing it.

03What gets unlocked at each price point

We keep an internal mapping of workflow types and the per-decision cost at which each becomes economical to fully automate, partially automate, or leave manual. As Graviton-class infrastructure improvements hit production, the threshold moves and the map updates.

The clients that captured the last cost drop were the ones whose workflows were ready to receive it. The ones who tried to redesign the workflow on the way down spent the cost difference on consulting and ended up where they would have been anyway. We expect the same pattern over the next eighteen months.

"Infrastructure papers are the leading indicator. Operating-model readiness is the lag indicator that decides who captures the gain."

How this maps to the work

We don't help clients build infrastructure. We help them be ready for it. Every six months a major lab publishes a paper like Graviton, and six to twelve months later the price of doing the work in production drops by a multiple. The clients that captured the previous drop were the ones whose workflows were already designed for the new economics.

Our work increasingly looks like this: take a workflow that isn't quite economical to fully automate today, redesign it so the human work and the model-eligible work are cleanly separable, and put it in a state where the next price drop is a configuration change rather than a re-engineering project.

Two engagements we run against this thesis.

None of these require a multi-year transformation. Each is scoped to land specific operating-model improvements with a measurable result.

01

Cost-per-decision modeling

We model your workflows in the unit that actually matters — cost-per-decision, not cost-per-token — at today's prices and at projected 18-month prices. The list of workflows that cross the threshold is your roadmap.

02

Separable workflow design

We rebuild workflows so the human steps and the model-eligible steps are cleanly separable. When prices drop, you move the dividing line — you don't redesign the workflow.

If this maps to what you're carrying — let's talk.

Most engagements start with a 30-minute conversation about the specific operating-model question on your desk this quarter.