This looks fantastic. Hopefully it may lead to some great things as I’ve always found the idea of exploiting the collective intelligence of the masses to be a terribly underused resource, and this reminds me of the game Foldit (and hopefully in the future will remind me of the wild success that that game had in the field of protein folding).
TinkerBird
This sounds like it would only work on a machine too dumb to be useful, and if it’s that dumb, you can switch it off yourself.
It doesn’t help with the convergent instrumental goal of neutralizing threats, because leaving a copy of yourself behind to kill all the humans allows you to be really sure that you’re switched off and won’t be switched on again.
I really appreciate these.
Why do some people think that alignment will be easy/easy enough?
Is there such thing as ‘aligned enough to help solve alignment research’?
I think there’s a lot we could learn from climate change activists. Having a tangible ‘bad guy’ would really help, so maybe we should be framing it more that way.
“The greedy corporations are gambling with our lives to line their pockets.”
“The governments are racing towards AI to win world domination, and Russia might win.”
“AI will put 99% of the population out of work forever and we’ll all starve.”
And a better way to frame the issue might be “Bad people using AI” as opposed to “AI will kill us”.
If anyone knows of any groups working towards a major public awareness campaign, please let the rest of us know about it. Or maybe we should start our own.
I’m with you on this. I think Yudkowsky was a lot better in this with his more serious tone, but even so, we need to look for better.
Popular scientific educators would be a place to start and I’ve thought about sending out a million emails to scientifically minded educators on YouTube, but even that doesn’t feel like the best solution to me.
The sort of people that are listened to are the more political types, so they I think are the people to reach out to. You might say they need to understand the science to talk about it, but I’d still put more weight on charisma vs. scientific authority.
Anyone have any ideas on how to get people like this on board?
As a note for Yudkowsky if he ever sees this and cares about the random gut feelings of strangers: after seeing this, I suspect the authoritative, stern strong leader tone of speaking will be much more effective than current approaches.
EDIT: missed a word
I’ve wanted something for AI alignment for ages like what the Foldit researchers created, where they turned protein folding into a puzzle game and the ordinary people online who played it wildly outperformed the researchers and algorithms purely by working together in vast numbers and combining their creative thinking.
I know it’s a lot to ask for with AI alignment, but still, if it’s possible, I’d put a lot of hope on it.
As someone who’s been pinning his hopes on a ‘survivable disaster’ to wake people up to the dangers, this is good news.
I doubt anything capable of destroying the world will come along significantly sooner than superintelligent AGI, and a world in which there are disasters due to AI feels like a world that is much more likely to survive compared to a world in which the whirling razorblades are invisible.
EDIT: “no fire alarm for AGI.” Oh I beg to differ, Mr. Yudkowsky. I beg to differ.
This confuses me too. I think Musk must be either smarter or a lot dumber than I thought he was yesterday, and sadly, dumber seems to be the way it usually goes.
That said, if this makes OpenAI go away to be replaced by a company run by someone who respects the dangers of AI, I’ll take it.
On the bright side… Nope, I’ve got nothing.
an AGI Risk Management Outreach Center with a clear cohesive message broadcast to the world
Something like this sounds like it could be a good idea. A way to make the most of those of us who are aware of the dangers and can buy the world time
Coordination will be the key. I wish we had more of it here on LW.
Like I say, not something I’d normally advocate, but no media stations have picked it up yet, and we might as well try whatever we can if we’re desperate enough.
We’ve never done a real media push but all indications are that people are ready to hear it.
I say we make a start on this ASAP.
What’s the consensus on David Shapiro and his heuristic imperatives design? He seems to consider it the best idea we’ve got for alignment and to be pretty optimistic about it, but I haven’t heard anyone else talking about it. Either I’m completely misunderstanding what he’s talking about, or he’s somehow found a way around all of the alignment problems.
Video of him explaining it here for reference, and thanks in advance:
With the TIME article, Bing’s AI’s aggression and Elon Musk’s every other tweet being about AI safety and STILL nothing happening in the media, I think it’s going to take something major to create the spark we need and then fan it into a real fire.
The best idea I’ve had so far is a letter writing campaign in the form of carefully worded emails from all of us here on LW to literally anyone with a platform, but I feel we’ve got to be able to do better than that.
Normally I wouldn’t advocate such a thing, but if we can’t convince people of the real big bag AI coming to get them, then the best thing we can do might be to create a fake and more tangible one.
“Misaligned AGI flying nanobots up our noses in 5 years”—no headlines
“AI takes control of Boston Dynamics robot and kills a man because it’s evil”—HEADLINES
Journalists are pretty much allowed to get away with straight up lying, so if anyone smarter than me could make a fake story that would actually resonate with people hit the news, it’d be something I’d put a lot of hope in.
I, for one, am looking forward to the the next public AI scares.
Same. I’m about to get into writing a lot of emails to a lot of influential public figures as part of a one man letter writing campaign in the hopes that at least one of them takes notice and says something publically about the problem of AI
PM’s are always open, my guy
but I haven’t seen anyone talk about this before.
You and me both. It feels like I’ve been the only one really trying to raise public awareness of this, and I would LOVE some help.
One thing I’m about to do is write the most convincing AI-could-kill-everyone email that I can that regulars Joes will easily understand and respect, and send that email out to anyone with a platform. YouTubers, TikTokers, people in government, journalists—anyone.
I’d really appreciate some help with this—both with writing the emails and sending them out. I’m hoping to turn it into a massive letter writing campaign.
Elon Musk alone has done a lot for AI safety awareness, and if, say, a popular YouTuber got one board, that alone could potentially make a small difference.
But if the current paradigm is not the final form of existentially dangerous AI, such research may not he particularly valuable.
I think we should figure out how to train puppies before we try to train wolves. It might turn out that very few principles carry over, but if they do, we’ll wish we delayed.
The only drawback I see to delaying is that it might cause people to take the issue less seriously than if powerful AI’s appear in their lives very suddenly.
I imagine it’s a sales tactic. Ask for $7 trillion, people assume you believe you’re worth that much, and if you’ve got such a high opinion of yourself, maybe you’re right…
In other news, I’m looking to sell a painting of mine for £2 million ;)