It is extremely difficult to gracefully put your finger on the scale of an LLM, to cause it to give answers it doesn’t ‘want’ to be giving. You will be caught.
IMO, this takeaway feels way too strong. It could just be that this wasn’t a very competent attempt. (And based on the system prompt we’ve seen, it sure looks like that.) How would we know if there were competent attempts we weren’t seeing?
IMO, this takeaway feels way too strong. It could just be that this wasn’t a very competent attempt. (And based on the system prompt we’ve seen, it sure looks like that.) How would we know if there were competent attempts we weren’t seeing?