Last Friday, a unified team from Stanford University and the University of Washington announced that they had trained a math and coding focused large language model that performs as well as OpenAI’s o1 and DeepSeek’s R1 reasoning models. It cost just $50 in cloud compute credits to build. The team reportedly used an off the shelf base model, then distilled Google’s Gemini 2.0 Flash Thinking Experimental model into it. The process of distilling AIs involves pulling the relevant...
![](/img/cat/257.jpg)