OpenAI targets "conversational" coding, not slow batch-style agents. Big latency wins: 80% faster roundtrip, 50% faster time-to-first-token. Runs on Cerebras WSE-3 chips for a latency-first Codex ...