I’m a software engineer that is interested in AI, futurism, space, and the big questions of life.
https://www.linkedin.com/in/jamessullivan092/
James Sullivan
Karma: 58
What Sentences Cause Alignment Faking?
Are we aligning the model or just its mask?
Playing Dumb: Detecting Sandbagging in Frontier LLMs via Consistency Checks
Jailbreaking Claude 4 and Other Frontier Language Models
How do AI agents work together when they can’t trust each other?
Of the people that wanted to go to a frontier lab, how many had a mentor that worked at a frontier lab? I assume that would make finding a role easier.
Would we really say that a human is a “narrow intelligence” when trying any new task until they sleep on it? I think the only thing that would meet the definition of “general intelligence” that this implies is something that generalize to all situations, no matter how foreign. By that definition, I’m not sure if general intelligence is possible.