That may be. I made my comment in reply to a previous version of Duncan’s comment (he edited after) which IIRC said specifically that I didn’t disclose conflicts of interest, and [some phrase I don’t recall, that I interpreted as, that I had said I wasn’t even trying to treat participants according to a certain standard of care]. That is false, and is the kind of thing I can know, so I commented with it. I don’t mean to imply that disclosing a conflict makes a person neutral on the subject, or that first person judgments of intent are reliable.
I did spend a good deal of time since 2020 discussing errors from my past CFAR stuff, and ways I and we may have harmed things we care about, but it was more along the lines of [doing stuff that is kinda common, but that messes with the epistemics of groups, and of individuals who sync with groups, a la Tsvi’s comment], not [skipping basic attempts to be honest, kind, respectful of expectations about confidentiality, mindful of a workshop participants’ well-being and of likely impacts on it, etc]. I agree with Tsvi that the epistemics-while-in-groups stuff is tricky for many groups, and is not much amenable to patching, which is more or less why this CFAR reboot took me five years, why I tested my guesses about how to do better in a couple smaller/easier group workshop contexts first, and why I’m calling these new CFAR workshops “pilots” and reserving the option to quit if the puzzles seem too hard for us. Tsvi I think remains mostly-pessimistic about my new plans, and I don’t, so there’s that.
Regardless:
Attending a workshop where groups of people think together about thinking, and practice new cognitive habits, while living together for five days and talking earnestly about a bunch of life stuff that one doesn’t normally discuss with near-strangers.… is indeed a somewhat risky activity, compared to eg attending a tennis-playing weekend or something else less introspective. Many will come out at least a little changed, and in ways that aren’t all deliberate. It’s worth knowing this, and thinking on whether one wants this.
A large portion of past CFAR participants said they were quite glad they came, including months and years afterward; and I suspect it was probably good for people on net (particularly people who passed through briefly, and retained independent cultural influences, I think?); but I also suspect there were a bunch of people who experienced subtle damage to parts of their skills and outlook and aesthetics in ways that were hard for them and us to track. And I expect some of that going forward, too, although I hope to make this less frequent via:
respecting individuals more deeply
having a plurality of frameworks that includes eg [falsification/feedback loops] and [pride in craftsmanship] and other stuff discussed in my OP rather than only [Bayesian updating + agentiness/goal-achievement]
having a “hobbyist convention” vibe where our guests are fellow hobbyists and can bring their own articulate or inarticulate frameworks
not myself being in “emergency mode” around AI risk (and being myself in something closer to a “geek out with people and try to be helpful and see where things go” mode, rather than in a super-goal-oriented mode), which I think should be better for not losing “peripheral vision” or “inarticulate-but-important bits of perception.”
One should not expect the CFAR alumni community to be only made of trustworthy carefully vetted people. We plan not to accept people who we suspect have unusually bad character; but I’m not that good at telling, and I don’t know that we as a team are, either. Also, there’s a question of false-negatives vs false-positives here, and I don’t plan to be maximally risk averse, although I do plan to try some; guests and alumni should keep in mind, when interacting with any future alumni community, that strangers vary in their trustworthiness.
I’m sometimes fairly skilled at seeing the gears inside peoples’ minds, especially when people try to open up to me; and when things are going well this can be useful to all parties. But I’ve harmed people via trying to dialog with bits of their mind that weren’t set up for navigating outside pressures, albeit mostly not mainline workshop participants, mostly in cases where I didn’t understand ways they were different from me and so moves that would’ve been okay on me were worse for them, and mostly in contact with peoples’ untempered urge to try to be relevant to AI safety which created a lot of fear/drive that might’ve been bumpy without me too (eg, folks who felt “I must continue to work at CFAR or MIRI or [wherever] or else my life won’t matter,” and so weren’t okay with the prospect of maybe-losing a job that most people would’ve quit because of the pain or difficulty of it). I do think I’m better at not causing harm in this way now (via chilling out in general, via somewhat better models of how some non-me people work, and via the slow accumulation of common sense), but whether other people have grounds to believe me about this varies.
Is the above enough bad that the world would be better off without CFAR re-opening and offering workshops? IMO, no. CFAR ran sixty-ish multi-day events from 2012-2020 with close to two thousand participants; some things went badly wrong, many things were meh compared to our hopes (our rationality curriculum is fairly cool, but didn’t feedback its way to superpowers); many things went gloriously right (new friendships; new businesses; more bay area rationality community in which many found families or other things they wanted; many alumni who tell me they learned, at the workshop, how to actively notice and change personal habits or parts of their lives that weren’t working for them). 2025 is somehow a time when many organizations and community structures have shut down; and I think there’s something sad about that and I don’t want CFAR’s to be one of them.
It seems good to me that people, including Duncan, want to tell their friends and family their views. (Also including people in top-level comments below this one who want to share positives; those are naturally a bit easier for me personally to enjoy.). A cacophony of people trying to share takes and info seems healthy to me, and healthier than a context where people-with-knowledge are pressured never to share negatives (or where people-with-knowledge who have positives are quiet about those in case CFAR is actually bad and they trick people).
I hope for relatively relaxed narrative structures both about CFAR and broadly, where peoples’ friends can help them make sense of whatever they are seeing and experiencing, and can help them get info they might want, in public view where sensible, without much all-or-nothingness. (I don’t mean from Duncan, whose honest take here is extremely negative, and who deserves to have that tracked; but from the mixture of everyone.)
That may be. I made my comment in reply to a previous version of Duncan’s comment (he edited after) which IIRC said specifically that I didn’t disclose conflicts of interest, and [some phrase I don’t recall, that I interpreted as, that I had said I wasn’t even trying to treat participants according to a certain standard of care]. That is false, and is the kind of thing I can know, so I commented with it. I don’t mean to imply that disclosing a conflict makes a person neutral on the subject, or that first person judgments of intent are reliable.
I did spend a good deal of time since 2020 discussing errors from my past CFAR stuff, and ways I and we may have harmed things we care about, but it was more along the lines of [doing stuff that is kinda common, but that messes with the epistemics of groups, and of individuals who sync with groups, a la Tsvi’s comment], not [skipping basic attempts to be honest, kind, respectful of expectations about confidentiality, mindful of a workshop participants’ well-being and of likely impacts on it, etc]. I agree with Tsvi that the epistemics-while-in-groups stuff is tricky for many groups, and is not much amenable to patching, which is more or less why this CFAR reboot took me five years, why I tested my guesses about how to do better in a couple smaller/easier group workshop contexts first, and why I’m calling these new CFAR workshops “pilots” and reserving the option to quit if the puzzles seem too hard for us. Tsvi I think remains mostly-pessimistic about my new plans, and I don’t, so there’s that.
Regardless:
Attending a workshop where groups of people think together about thinking, and practice new cognitive habits, while living together for five days and talking earnestly about a bunch of life stuff that one doesn’t normally discuss with near-strangers.… is indeed a somewhat risky activity, compared to eg attending a tennis-playing weekend or something else less introspective. Many will come out at least a little changed, and in ways that aren’t all deliberate. It’s worth knowing this, and thinking on whether one wants this.
A large portion of past CFAR participants said they were quite glad they came, including months and years afterward; and I suspect it was probably good for people on net (particularly people who passed through briefly, and retained independent cultural influences, I think?); but I also suspect there were a bunch of people who experienced subtle damage to parts of their skills and outlook and aesthetics in ways that were hard for them and us to track. And I expect some of that going forward, too, although I hope to make this less frequent via:
respecting individuals more deeply
having a plurality of frameworks that includes eg [falsification/feedback loops] and [pride in craftsmanship] and other stuff discussed in my OP rather than only [Bayesian updating + agentiness/goal-achievement]
having a “hobbyist convention” vibe where our guests are fellow hobbyists and can bring their own articulate or inarticulate frameworks
not myself being in “emergency mode” around AI risk (and being myself in something closer to a “geek out with people and try to be helpful and see where things go” mode, rather than in a super-goal-oriented mode), which I think should be better for not losing “peripheral vision” or “inarticulate-but-important bits of perception.”
One should not expect the CFAR alumni community to be only made of trustworthy carefully vetted people. We plan not to accept people who we suspect have unusually bad character; but I’m not that good at telling, and I don’t know that we as a team are, either. Also, there’s a question of false-negatives vs false-positives here, and I don’t plan to be maximally risk averse, although I do plan to try some; guests and alumni should keep in mind, when interacting with any future alumni community, that strangers vary in their trustworthiness.
I’m sometimes fairly skilled at seeing the gears inside peoples’ minds, especially when people try to open up to me; and when things are going well this can be useful to all parties. But I’ve harmed people via trying to dialog with bits of their mind that weren’t set up for navigating outside pressures, albeit mostly not mainline workshop participants, mostly in cases where I didn’t understand ways they were different from me and so moves that would’ve been okay on me were worse for them, and mostly in contact with peoples’ untempered urge to try to be relevant to AI safety which created a lot of fear/drive that might’ve been bumpy without me too (eg, folks who felt “I must continue to work at CFAR or MIRI or [wherever] or else my life won’t matter,” and so weren’t okay with the prospect of maybe-losing a job that most people would’ve quit because of the pain or difficulty of it). I do think I’m better at not causing harm in this way now (via chilling out in general, via somewhat better models of how some non-me people work, and via the slow accumulation of common sense), but whether other people have grounds to believe me about this varies.
Is the above enough bad that the world would be better off without CFAR re-opening and offering workshops? IMO, no. CFAR ran sixty-ish multi-day events from 2012-2020 with close to two thousand participants; some things went badly wrong, many things were meh compared to our hopes (our rationality curriculum is fairly cool, but didn’t feedback its way to superpowers); many things went gloriously right (new friendships; new businesses; more bay area rationality community in which many found families or other things they wanted; many alumni who tell me they learned, at the workshop, how to actively notice and change personal habits or parts of their lives that weren’t working for them). 2025 is somehow a time when many organizations and community structures have shut down; and I think there’s something sad about that and I don’t want CFAR’s to be one of them.
It seems good to me that people, including Duncan, want to tell their friends and family their views. (Also including people in top-level comments below this one who want to share positives; those are naturally a bit easier for me personally to enjoy.). A cacophony of people trying to share takes and info seems healthy to me, and healthier than a context where people-with-knowledge are pressured never to share negatives (or where people-with-knowledge who have positives are quiet about those in case CFAR is actually bad and they trick people).
I hope for relatively relaxed narrative structures both about CFAR and broadly, where peoples’ friends can help them make sense of whatever they are seeing and experiencing, and can help them get info they might want, in public view where sensible, without much all-or-nothingness. (I don’t mean from Duncan, whose honest take here is extremely negative, and who deserves to have that tracked; but from the mixture of everyone.)