Posts
Comments
Reposted: https://www.facebook.com/groups/159388659401670/permalink/165329602140909
How would you feel about reposting this in EAs for Political Tolerance (https://www.facebook.com/groups/159388659401670) ? I'd also be happy to repost it for you if you'd prefer.
I'm honestly a bit flummoxed here. Why would contributing to a Facebook group explicitly aligned with one side of this dispute help avoid a split?
The group is still new, so it's still unclear exactly how it'll turn out. But I don't think that's a completely accurate way of characterisating the group. I expect that there are two main strands of thought within the group - some see themselves as fighting against woke tendencies, whilst others are more focused on peace-making and want to avoid taking a side.
"On the other hand, we've had quite a bit of anti-cancel-culture stuff on the Forum lately. There's been much more of that than of pro-SJ/pro-DEI content, and it's generally got much higher karma. I think the message that the subset of EA that is highly active on the Forum generally disapproves of cancel culture has been made pretty clearly"
Perhaps. However, this post makes specific claims about ACE. And even though these claims have been discussed somewhat informally on Facebook, this post provides a far more solid writeup. So it does seem to be making a signficantly new contribution to the discussion and not just rewarming leftovers.
It would have been better if Hypatia had emailed the organisation ahead of time. However, I believe ACE staff members might have already commented on some of these issues (correct me if I'm wrong). And it's more of a good practise than something than a strict requirement - I totally understand the urge to just get something out of there.
"I'm sceptical that further content in this vein will have the desired effect on EA and EA-adjacent groups and individuals who are less active on the Forum, other than to alienate them and promote a split in the movement, while also exposing EA to substantial PR risk"
On the contrary, now that this has been written up on the forum it gives people something to link to. So forum posts aren't just read by people who regularly read the forum. In any case, this kind of high quality write-up is unlikely to have a significnat effect on alienating people compared to some of the lower quality discussions on these topics that occur in person or on Facebook. So, from my perspective it doesn't really make any sense to be focusing on this post. If you want to avoid a split in the movement, I'd like to encourage you to join the Effective Altruists for Political Tolerance Facebook group and contribute there.
I would also suggest worrying less about PR risks. People who want to attack EA can already go around shouting about 'techno-capitalists', 'overwhelmingly white straight males', 'AI alarmists', ect. If someone wants to find something negative, they'll find something negative.
Part of my model is that there is decreasing marginal utility as you invest more effort in one form of outreach, so there can be significant benefit in investing some resources in investing small amounts of resources in alternate forms of outreach.
I hope you find finding to pay someone to organise this as I suspect this program could be extremely impactul.
I would also love to see some amount of prize money funded for this. I wouldn't be surprised if a relatively small amount of money by philanthropic standards could tempt more of the top debaters to enter.
I actually found the Facebook group very difficult to search for - link is here.
Making a Wiki successful is always about seeding content. There's a lot of past content that could be copied over and updated, but it's not pleasant work, so it's good that Pablo has a grant.
As an addendum: First, suppose you compare a group of random people from the same demographic to a random groups of people from different demographics. Next suppose you compare a group of random lawyers to a group of random laywers of different demogaphics. I would suggest that in the second case the increase in diversity from adding demographic diversity would be significantly reduced as the bar to becoming a lawyer would filter out a lot of diversity of experiences from the first case. For example, a greater proportion of African Americans experience poverty than the general population, but the difference among those who become laywers would be much less.
"They were founded under the premise that conservative viewpoints are underrepresented in scientific discourse" - that's definitely a possibility, although I suspect that for research into underrepresented groups in general almost all research will have been conducted by people withn strong pre-existing beliefs about whether or not such a group is underrepresented.
I think there's value in considering people's possible psychological motivations, but I find it more helpful to consider these for all parties. In such a conversation, the rich could very well be afraid of losing their privilege and the poor could very well be jealous or resentful.
It was a general comment how this lens is often applied in practise, even though this isn't the only possible way for it to be applied.
"As we cannot measure the diversity of perspectives of a person directly, our best proxy for it is demographic diversity"
Demographic diversity is a useful proxy and may add something additional even if we did have diversity of general philosophy. However, we can measure diversity of perspectives directly, ie. by running surveys like Heterodox Academy has.
"The answer here is that objectivity is not something that a single person has, but that objectivity is a social achievement of a diverse community"
Feminism offers some valuable lens, but I feel it often leads to a hyperfocus on the underprivileged. Suppose we're discussing raising taxes on the rich, it might be useful to have a rich guy in a room. They might share some useful perspectives like, "It won't change the behaviour of my friends one bit. Most of us won't even notice. Our accountants handled our taxes, so we have no idea how much we're paying" or "If the California tax law passes, I'm headed to Texas". They might lie, but that's true of everyone. They might be biased, but the poor are likely to be biased as well.
I'm not claiming this is equally important as representing the perspectives of the poor, just that we shouldn't be hyperfocused.
I also think charity science might have tried getting people to pledge in their wills.
Yeah, hopefully at some point I find time to make another post, linking to various aspects of what I'd define as the community. I guess who is in or not is not well-defined as it's not really a single community. Rather, it's a bunch of groups with similar kinds of people who seem to be talking to each and talking about similar kinds of things, most of whom I think would agree that they're doing something like sensemaking.
Regarding your second question, if you head over to the Stoa or listen to Both/And, you'll see people from across the spectrum, although not really many strong social justice proponents. I suppose my suspicion is mainly driven by the intuition that ending the culture wars requires a movement with positive content of its own and not merely a negative critique as Quillette and (to a lesser degree) Persuasion seem to do. People need a reason to join apart from simply being sick of the culture wars.
Yeah, I agree that there would be significant benefits to trying to set up another academic research institute at a university more focused on economics.
This is full, but it's worth getting people to subscribe for the future
Hmm... often I think it is nice to have a standard term for a phenomenon so that people don't have to figure out how to express a certain concept each time and then hope that everyone else can follow. Language also has the advantage that insofar as we convince people to adopt our language, we draw them into our worldview.
This should really be a Wiki page instead since these lists (I even made one myself in the past) always become outdated.
This is a really challenging situation - I could honestly see myself leaning either way on this kind of scenario. I used to lean a lot more towards saying whatever I thought was true and ignoring the consequences, but lately I've been thinking that it's important to pick your battles.
I think the key sentence is this one - "On many subjects EAs rightfully attempt to adopt a nuanced opinion, carefully and neutrally comparing the pros and cons, and only in the conclusion adopting a tentative, highly hedged, extremely provisional stance. Alas, this is not such a subject."
What seems more important to me is not necessarily these kinds of edge cases, but that we talk openly about the threat potentially posed. Replacing the talk with a discussion about cancel culture instead seems like it could have been a brilliant Jiu Jitsu move. I'm actually much more worried about what's been going on with ACE than anything else.
Thanks, that was useful. I didn't realise that his argument involved 1+2 and not just 1 by itself. That said, if the hinge of history was some point in the past, then that doesn't affect our decisions as we can't invest in the past. And perhaps it's a less extraordinary coincidence that the forward-looking hinge of history (where we restrict the time period from now until the end of humanity) could be now, especially if in the average case we don't expect history to go on much longer.
I've never found Will's objections to the hinge of history argument persuasive. Convincing me that there was a greater potential impact in past times than I thought, ie. that it would have been very influential to prevent the rise of Christianity, shouldn't make me disbelieve that arguments that AI or bio risks are likely to lead to catastrophe in the next few decades if we don't do anything about it. But maybe I just need to reread the argument again.
DNA engineering has some positive points, but imagine the power that having significant control its citizens personalities would give the government. That shouldn't be underestimated.
The real hinge here is how much we should expect the future to be a continuation of the past and how much we update based on our best predictions. Given what we know about about existential risk and the likelihood that AI will dramatically change our economy, I don't think this idea makes sense in the current context.
I agree that such a system would be terrifying. But I worry that its absence would be even more terrifying. Limited surveillance systems work decently for gun control, but when we get to the stage where someone can kill tens of thousands or even millions instead of a hundred I suspect it'll break down.
Thanks for posting such a detailed answer!
It's great to hear that you are setting this up. However, the current post seems light on details. Why are these areas of particular interest? What kind of commitment are you hoping for from participants?
I think people are quite reasonably deciding that this post isn't worth taking the time to engage with. I'll just make three points even though I could make more:
"A good rule of thumb might be that when InfoWars takes your side, you probably ought to do some self-reflection on whether the path your community is on is the path to a better world." - Reversed Stupidity is Not Intelligence
"In response, the Slate Star Codex community basically proceeded to harass and threaten to dox both the editor and journalist writing the article. Multiple individuals threatened to release their addresses, or explicitly threatened them with violence." - The author is completely ignoring the fact that Scott Alexander specifically told people to be nice, not to take it out on them and didn't name the journalist. This seems to suggest that the author isn't even trying to be fair.
"I have nothing to say to you — other people have demonstrated this point more clearly elsewhere" - I'm not going to claim that such differences exist, but if the author isn't open to dialog on one claim, it's reasonable to infer that they mightn't be open to dialog on other claims even if they are completely unrelated.
Quite simply this is a low quality post and "I'm going to write a low quality post on topic X and you have to engage with me because topic X is important regardless of the quality" just gives a free pass on low quality content. But doesn't it spur discussion? I've actually found that most often low quality posts don't even provide the claimed benefit. They don't change people's minds and tend to lead to low quality discussion.
It's worth remembering though that when people who paid for the book are much more likely to have read it
What do you think about the fact that many in the field are pretty open that they are pursuing enquiry on how to achieve an ideology and not neutral enquiry (using lines like all fields are ideological whether they know it or not)?
"It's a bit concerning that the community level of knowledge of the bodies of work that deal with these issues is just average" - I do think there are valuable lessons to be drawn from the literature, unfortunately a) lots of the work is low quality or under-evidenced b) discussion of these issues often ends up being highly divisive, whilst not changing many people's minds
"If the self-oriented reasons for action leave it largely underdetermined how personal flourishing would look like" - If we accept pleasure and pain, then we can evaluate other actions in high likely they are to lead to pleasure/pain in the long term, so I don't see how actions are underdetermined.
I'm surprised that you put moral realism on the same tier as self-oriented reasons for action. It would seem much more astounding to claim that pain and pleasure are neither good nor bad *for me*, then to claim that there's no objective stance by which others should consider my pain good or bad. The Pascal's wager argument is also much stronger too.
I think percentages are misleading. In terms of influencing demographic X, what matters isn't so much how many people of demographic X there are in these organisations, but how well-respected they are.
I'm generally against evaluating diversity programs by how much diversity they create. It's definitely a relevant metric, but we don't evaluate AMF by how many bednets they hand out, but the impact of these bednets.
I guess that seems so far off that I wasn't focusing on it. I'm more interesting in how to establish a working impact purchase in the mean time.
"These people where encouraging me to use my last savings to retrain to a risky career, but putting in their money was out of the question" - Yeah, I'm sorry you had that experience, that seems unpleasant. Maybe they didn't understand the financial precariousness of your situation? Like many EAs wouldn't find it hard to land a cushy job in normal times and likely have parents who'd bail them out worst comes to worst and might have assumed that you were in the same position without realising they were making an assumption?
Or if you think an outcome was mostly bad luck, fund them more than just impact purchase
Yeah, luck is another argument I considered covering, but didn't get into. Sometimes the impact of a project is just a matter of being at the right place in the right time. Of course, it's hard to tell; to a certain extent people make their own luck.
But in most cases I would suggest straight up impact purchase, because anything else is is really hard and you'll probably get the adjustments wrong.
I guess this would be a key point where we differ. I haven't thought deeply about this, but my intuition would be that adjustments would greatly improve impact. For example, a small project extremely competently implemented and a big project poorly implemented might have the exact same impact, but the former would be a stronger signal.
I wrote up my thoughts in this post: Making Impact Purchases Viable. Briefly, I argue that:
- Restrictions on sellers are necessary to fix and imbalance between the size of the buyer's market and seller's market
- Restrictions are also important for providing sufficient incentive for a reasonable proportion of the impact to be counterfactual
- Another issue I don't have a solution to is the potential of impact purchases to lead to bitterness or demotivate people
That line of reasoning also suggests that EA orgs should ask each of their employees whether they will work for free if they don't get a salary; and refuse to pay a salary to employees who answer "yes".
Maybe, but I imagine that the number of people who'd work just as hard long-term would be about zero, so more of the impact would be counterfactual.
Even though fair trade is ineffective on an individual level, it may be effective on a collective level because enough people find it appealing for broad adoption. Deciding to ignore it weakens any attempt to establish buying fairtrade as a society.
EAs don't arise out of a vacuum, but out of society. If society is doing well, then EAs are more likely to do well too and hence to have more impact. So by not donating to a local charity, you are refusing to invest in the society that provided you the chance to have an impact in the first place.
Not saying you should donate locally or buy fair trade, just pointing out one worry with ignoring them.
Thanks so much for writing this. I've had similar worries regarding local charity and things like fair trade for a while.
It's hard to do a summary without encouraging people to read the summary instead of the article.
My understanding from looking briefly was that Drawdown focused on total reduction potential, not cost-effectiveness
These diagrams look really useful for encouraging people to map out potential paths to existential risk and potential interventions more carefully!
I suspect that part of the reason why this is happened is that the EA community is closely associated with the rationality community, so it's often easiest just to have the personal development discussion online over there. Plus another reason that people mightn't feel a need for it online is that lots of discussion of personal development occurs informally at meetups.
What's Good Growth?
Yeah, EA is likely less compelling when this is defined as feeling motivating/interesting to the average person at the moment, although it is hard to judge since EA hasn't been around for anywhere near as long. Nonetheless, many of the issues EAs care about seem way too weird for the average person, then again if you look at feminism, a lot of the ideas were only ever present in an overly academic form. Part of the reason why they are so influential now is that they have filtered down into the general population in a simpler form (such as "girl power", "feeling good, rationality bad"). Plus social justice is more likely to benefit the people supporting it in the here and now than EA which focuses more on other countries, other species and other times which is always a tough sell.
SJ is an extremely inclusive movement (basically by definition)
I'm generally wary of argument by definition. Indeed, SJ is very inclusive to members of a racial minority or those who are LGBTI, but is very much not when it comes to ideological diversity. And some strands can be very unwelcoming to members of majorities. So it's much more complex than that.
"There are definitely many who see these more in the movement/tribe sense" - For modern social justice this tends to focus on who is a good or bad person, while for EA this tends to focus more on who to trust. (There's a less dominant strand of thought within social justice that says we shouldn't blame individuals for systematic issues, but it's relatively rare). EA makes some efforts towards being anti-tribal, while social justice is less worried about the downsides of being tribal.
Greater knowledge of psychology would be powerful, but why should we expect the sign to be positive, instead of say making the world worse by improving propaganda and marketing?
Why is Leverage working on psychology? What is it hoping to accomplish?