I actually told the most hippie human on my list (spending months on rainbow gatherings-level hippie) that she’s on it. To my surprise, she felt unambiguously flattered. Seems like the people who know me trust that I can be intentional without being objectifying. :)
Severin T. Seehrich
Advice for newly busy people
AGI safety field building projects I’d like to see
Advice for interacting with busy people
Guidelines for cold messaging people
The Dunbar Playbook: A CRM system for your friends
AGI Safety Needs People With All Skillsets!
EA might systematically generate a scarcity mindset that produces low-integrity actors
Tend to your clarity, not your confusion
The Conversations We Make Space For
Community building: Lessons from ten years of facilitation experience
Sequence opener: Jordan Harbinger’s 6 minute networking
Hah, this makes a lot of sense. Thanks!
An addition to that: If we look through the goggles of Sara Ness’ Relating Languages, the rationalist style of doing conversations is at the far end of the internal-focusing dialects Debater/Chronicler/Scientist. In my experience, more gooey communities have way more Banterer/Bard/Spaceholder-heavy types of interactions, which focus more on peoples’ needs in the situation than on forming and communicating true beliefs. People don’t necessarily know which dialects they speak themselves, because their way of interacting just feels normal to them, and everyone else weird. It’s hard to learn speak in dialects that are not your natural default. For example, I didn’t even notice myself slipping into Bard/Banterer during writing this post, but in hindsight it’s fairly obvious how it digresses from the LessWrong language game.
I think the LW-way is ideal for its purpose, but I’m realizing that there’s a whole lot of tacit knowledge and implicit norms involved in understanding and doing it. This strong selection for a particular style of communication may be responsible for a significant chunk of the difficulty I’m perceiving in interfacing between the rationalist and other memeplexes. In both directions, both for the rationalist community learning from other memeplexes, and for useful memes getting from rationalist circles into the outside world.
Highly depends on your role and personality I guess.
As a community builder and someone pretty high on extraversion, I’m generally happy to add more people to my loose network. If there’s just a bit of overlap between my and a stranger’s interests, I expect there to be a far higher upside than downside risk to us knowing that the other exists and what they work on. Of course, I may change my opinion on this over time while my time becomes more valuable and my loose network larger.
Any generalizable rules you can think of about whom better not to cold message at all?
I did not recommend any particular intervention in my post. I just tried to explain some part of my understanding of how new psycho- and social technologies are generated, and what conclusions I draw from that.
If you expect most if not all established therapeutic interventions to not survive the replication crisis—what would you consider sufficient evidence for using or suggesting a certain intervention?
For example, a friend of mine felt blue today and I sent them a video of an animated dancing seal without extensively googling for meta-analyses on the effect of cute seal videos on peoples’ moods beforehand. Would you say I had sufficient evidence to assume that doing so is better than not doing so? Or did I commit epistemic sin in making that decision? This is an honest question, because I don’t yet get your point.
Agreed. But sitting around and sulking is a bummer, so I rather keep learning, exploring, and sometimes finding things that work for me.
So, in other words—I am wrong, hippies are wrong, and most if not all therapies that look so far like they are backed by evidence are likely wrong, too.
Who or what do you suggest we turn to for fixing our stuff?
Agreed—I added the 7th point to the list now to account for this.
I’m not in London, but aisafety.community (the afaik most comprehensive and way too unknown resource on AI safety communities) suggests the London AI Safety Hub. There are some remote alignment communities mentioned on aisafety.community as well. You might want to consider them as fallback options, but probably already know most if not all of them.
Let me know if that’s at all helpful.
This is one of the points I’m less sure about because often enough, the rest of the message will implicitly answer it. In addition, what to include is highly dependent on context and who you are writing to.
Two very general recommendations:
- Something that helps the other person gauge how long the inferential distances between you two are, so that communication can be as quick as possible and as thorough as necessary.
- Something that helps them gauge your level of seniority. It’s unfortunate but true that the time of people a couple levels of seniority above your own is extremely valuable. For example, it would hardly make sense for a Nick Bostrom to make time for helping a bright-but-not-Einstein-level high school student he never met decide which minor to choose in university. If people can’t gauge your level of seniority, they might misjudge whether they are the right person for you to talk to, and then you might end up in a conversation that is extremely awkward and a waste of time for either side.
Some examples:
- “Hi! I’m xyz, Ops lead at Linear.”
- “Hi! I’m a computer science undergrad at Atlantis University and a long-time lurker on LessTrue.”
- …