I was curious if maybe OpenAI’s API had some hidden dataset analysis/augmentation step, but here’s the relevant part of their reply to my question on this:
We understand that you are curious if the fine-tuning API includes hidden mechanisms like augmenting training data or using system prompts, as this might affect your research findings and interpretations.
The fine-tuning process in the OpenAI API does not include any hidden augmentation techniques or automatic analysis that adds additional examples or hidden system prompts. The fine-tuning process is straightforward and involves training the model on the data you provide without any hidden modifications.
This refers only to the regular old finetuning, for 4o, and not to the fancy new RL finetuning for o1 that they recently opened up to alpha users, right?
I was curious if maybe OpenAI’s API had some hidden dataset analysis/augmentation step, but here’s the relevant part of their reply to my question on this:
This refers only to the regular old finetuning, for 4o, and not to the fancy new RL finetuning for o1 that they recently opened up to alpha users, right?
Correct.
Edit: I just realized you may have meant one of two things:
The post above was with regular 4o fine-tuning.
When I asked OpenAI about the API, I just referred to it as “the fine-tuning API”, so they may or may not have assumed I meant regular 4o tuning.