I think you just do good research, and let it percolate through the intellectual environment. It might be helpful to bug org people to look at safety research, but probably not a good idea to bug them to look at yours specifically.
I am curious why you expect AGI will not be a scaffolded LLM but will be the result of self-play and massive training runs. I expect both.
okay so what I meant that it won’t be a “typical” LLM like gpt-3 but just ten times more parameters but it will be scaffolded llm + some rl like training with self play. Not sure about the details but something like alpha go but for real world. Which I think agrees with what you said.
I think you just do good research, and let it percolate through the intellectual environment. It might be helpful to bug org people to look at safety research, but probably not a good idea to bug them to look at yours specifically.
I am curious why you expect AGI will not be a scaffolded LLM but will be the result of self-play and massive training runs. I expect both.
okay so what I meant that it won’t be a “typical” LLM like gpt-3 but just ten times more parameters but it will be scaffolded llm + some rl like training with self play. Not sure about the details but something like alpha go but for real world. Which I think agrees with what you said.