Making a research platform for AI Alignment at https://ai-plans.com/
Come critique AI Alignment plans and get feedback on your alignment plan!
Iknownothing
AI-Plans.com—a contributable compendium
Review of Alignment Plan Critiques- December AI-Plans Critique-a-Thon Results
An Ignorant View on Ineffectiveness of AI Safety
Critique-a-Thon of AI Alignment Plans
Even briefer summary of ai-plans.com
Brief summary of ai-plans.com
No. Humans are not large networks that can be quickly and easily controlled. Among many, many other differences.
AI-Plans.com 10-day Critique-a-Thon
When I say media, I mean social media, movies, videos, books etc- any type of recording or something that you believe you’re using as entertainment.
I’m trying this myself. Done singular days before, sometimes 2 or 3 days, but failed to keep it consistent. I did find that when I did it, my work output was far higher and greater quality, I had a much better sleeping schedule and was generally in a much more enjoyable mood.
I also ended up spending more time with friends and family, meeting new people, trying interesting things, spending time outdoors, etc.
This time I’m building up to it- starting with 1 media free hour a day, then 2 hours, then 3, etc.
I think building up to it will let me build new habits which will stick more.
A challenge for folks interested: spend 2 weeks without media based entertainment.
It’s not directly about AGI, no. But it could be a way to change a skeptic’s mind about AI risk. Which could be useful if they’re a regulator/politician.
This plan originated from the idea of trying to have a hackathon to disprove alignment plans. I’m still very interested in that!
Looking for judges for critiques of Alignment Plans
AI Law-a-Thon
You disagree with doomerism as a mindset, or factual likelihood? Or both?
I think doomerism as a mindset isn’t great, but in terms of likelihood, there are ~3 things likely to kill humanity atm. AI being the first.
[Question] Specific Arguments against open source LLMs?
Simple alignment plan that maybe works
I’m generally disincentivized to post or put effort into a post from the system where someone can just heavily downvote my post, without even giving a reason.
“CESI’s Artificial Intelligence Standardization White Paper released in 2018 states
that “AI systems that have a direct impact on the safety of humanity and the safety of life,
and may constitute threats to humans” must be regulated and assessed, suggesting a broad
threat perception (Section 4.5.7).42 In addition, a TC260 white paper released in 2019 on AI
safety/security worries that “emergence” (涌现性) by AI algorithms can exacerbate the
black box effect and “autonomy” can lead to algorithmic “self-improvement” (Section
3.2.1.3).43”
From https://concordia-consulting.com/wp-content/uploads/2023/10/State-of-AI-Safety-in-China.pdf