First model from the Cerebras partnership delivers 1000+ tokens per second—10-20x faster than typical GPT models. Powered by Wafer Scale Engine 3, designed for real-time coding with 80% reduced roundtrip overhead. ChatGPT Pro users can access the research preview starting today.