NHacker Next
  • new
  • past
  • show
  • ask
  • show
  • jobs
  • submit
Sky-T1: Train your own O1 preview model within $450 (novasky-ai.github.io)
elashri 5 days ago [-]
> We initially trained a 32B model using 3–4K math problems from the Numina dataset (provided by STILL-2), achieving a significant improvement in AIME24 accuracy from 16.7% to 43.3%. However, when we incorporated coding data generated from the APPs dataset into the training process, AIME24 accuracy dropped to 36.7%. We hypothesize that this decline is due to the distinct reasoning approaches required for math and coding tasks.

This is interesting. For large models that were trained on much more data. I wonder if o1 is trained in a different way that GPT-4o. Do they only rely on synthetic data (plus some hand crafted datasets). But then how would O1 knows a lot of facts like GPT-4o indicating that these were in the training.

Can someone with more understanding and knowledge weight on this?

thot_experiment 4 days ago [-]
They finetuned QwQ to perform well on a benchmark. For the past two years there has been a constant stream of "X fine-tune beats Y closed model on Z benchmark". This isn't interesting and has never been interesting, see Goodhart's Law. If you're actually using local models day to day you will quickly find that finetunes are almost universally a waste of time. Even when it comes to something like smutty roleplay gaslighting a model can often lead to more interesting and consistent results because finetunes are basically always overfit to the training data.
zamadatix 5 days ago [-]
Fine tune an existing model. Training such a model so cheaply would've been nuts.
ipsum2 5 days ago [-]
For math only.
Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact
Rendered at 07:33:11 GMT+0000 (Coordinated Universal Time) with Vercel.