How can you tell those are 1. jailbroken 2. open-weights models? I use LLMs for coding a lot but I cant even tell apart claude and chatgpt confidently.
People are posting themselves jailbreaking or even releasing tools to jailbreak models. I meant this is done explicitly and celebrated, not that I am implicitly noticing it.
How can you tell those are 1. jailbroken 2. open-weights models? I use LLMs for coding a lot but I cant even tell apart claude and chatgpt confidently.
People are posting themselves jailbreaking or even releasing tools to jailbreak models. I meant this is done explicitly and celebrated, not that I am implicitly noticing it.