If they are open-source, then doesn’t it mean that anyone can check how the models’ alignment is influenced by training or adding noise? Or does it mean that anyone can repeat the training methods?
Generally the releases of the “open source” models release the inference code and the weights, but not the exact training data, and often not information about training setup. (for instance, Deepseek has done a pile of hacking on how to get the most out of their H800s, which is private)
If they are open-source, then doesn’t it mean that anyone can check how the models’ alignment is influenced by training or adding noise? Or does it mean that anyone can repeat the training methods?
Generally the releases of the “open source” models release the inference code and the weights, but not the exact training data, and often not information about training setup. (for instance, Deepseek has done a pile of hacking on how to get the most out of their H800s, which is private)