Comment by JoshYou on What Makes Outreach to Progressives Hard · 2021-03-17T01:39:59.116Z · EA · GW

Longtermism isn't just AI risk, but concern with AI-risk is associated with a Elon Musk-technofuturist-technolibertarian-Silicon Valley idea cluster. Many progressives dislike some or all of those things and will judge AI alignment negatively as a result.

Comment by JoshYou on Ask Rethink Priorities Anything (AMA) · 2020-12-14T17:18:26.125Z · EA · GW

How's having two executive directors going?

Comment by JoshYou on Ask Rethink Priorities Anything (AMA) · 2020-12-14T17:16:40.318Z · EA · GW

How do you decide how to allocate research time between cause areas (e.g. animals vs x-risk)?

Comment by JoshYou on Some thoughts on the EA Munich // Robin Hanson incident · 2020-09-09T02:40:10.294Z · EA · GW

My description was based on Buck's correction (I don't have any first-hand knowledge). I think a few white nationalists congregated at Leverage, not that most Leverage employees are white nationalists, which I don't believe. I don't mean to imply anything stronger than what Buck claimed about Leverage.

I invoked white nationalists not as a hypothetical representative of ideologies I don't like but quite deliberately, because they literally exist in substantial numbers in EA-adjacent online spaces and they could view EA as fertile ground if the EA community had different moderation and discursive norms. (Edited to avoid potential collateral reputational damage) I think the neo-reactionary community and their adjacency to rationalist networks are a clear example.

Comment by JoshYou on Some thoughts on the EA Munich // Robin Hanson incident · 2020-09-09T02:34:07.838Z · EA · GW

I also agree that it's ridiculous when left-wingers smear everyone on the right as Nazis, white nationalists, whatever. I'm not talking about conservatives, or the "IDW", or people who don't like the BLM movement or think racism is no big deal. I'd be quite happy for more right-of-center folks to join EA. I do mean literal white nationalists (like on par with the views in Jonah Bennett's leaked emails. I don't think his defense is credible at all, by the way).

I don't think it's accurate to see white nationalists in online communities as just the right tail that develops organically from a wide distribution of political views. White nationalists are more organized than that and have their own social networks (precisely because they're not just really conservative conservatives). Regular conservatives outnumber white nationalists by orders of magnitude in the general public, but I don't think that implies that white nationalists will be virtually non-existent in a space just because the majority are left of center.

Comment by JoshYou on Some thoughts on the EA Munich // Robin Hanson incident · 2020-09-08T22:39:18.370Z · EA · GW

We've already seen white nationalists congregate in some EA-adjacent spaces. My impression is that (especially online) spaces that don't moderate away or at least discourage such views will tend to attract them - it's not the pattern of activity you'd see if white nationalists randomly bounce around places or people organically arrive at those views. I think this is quite dangerous for epistemic norms, because white nationalist/supremacist views are very incorrect and deter large swaths of potential participants and also people with those views routinely argue in bad faith by hiding how extreme their actual opinions are while surreptitiously promoting the extreme version. It's also in my view a fairly clear and present danger to EA given that there are other communities with some white nationalist presence that are quite socially close to EA.

Comment by JoshYou on If a poverty alleviation intervention has a positive ROI, (why) isn't anyone lending money for them? · 2020-08-26T21:03:11.936Z · EA · GW

This is essentially the premise of microfinance, right?

Comment by JoshYou on Will Three Gorges Dam Collapse And Kill Millions? · 2020-07-26T15:01:35.638Z · EA · GW

From what I understand, since Three Gorges is a gravity dam, meaning it uses the weight of the dam to hold back water rather than its tensile strength, a failure or collapse would not necessarily be catastrophic one. So if some portion falls, the rest will stay standing. That means there's a distribution of severity within failures/collapses, it's not just a binary outcome.

Comment by JoshYou on Longtermism ⋂ Twitter · 2020-06-16T16:40:01.436Z · EA · GW

To me it feels easier to participate in discussions on Twitter than on (e.g.) the EA Forum, even though you're allowed to post a forum comment with fewer than 280 characters. This makes me a little worried that people feel intimidated about offering "quick takes" here because most comments are pretty long. I think people should feel free to offer feedback more detailed than an upvote/downvote without investing a lot of time in a long comment.

Comment by JoshYou on 80,000 Episode Re: Size of Community · 2020-06-16T15:57:42.325Z · EA · GW

Not from the podcast but here's a talk Rob gave in 2015 about potential arguments against growing the EA community:

Comment by JoshYou on Notes on how a recession might impact giving and EA · 2020-03-16T23:12:40.172Z · EA · GW

EAs are probably more likely than the general public to keep money they intend to donate invested in stocks, since that's a pretty common bit of financial advice floating around the community. So the large drop in stock prices in the past few weeks (and possible future drops) may affect EA giving more than giving as a whole.

Comment by JoshYou on AMA: Rob Mather, founder and CEO of the Against Malaria Foundation · 2020-01-27T16:46:32.674Z · EA · GW

How far do you think we are from completely filling the need for malaria nets, and what are the barriers left to achieving that goal?

Comment by JoshYou on I'm Cullen O'Keefe, a Policy Researcher at OpenAI, AMA · 2020-01-11T18:08:09.397Z · EA · GW

What are your high-level goals for improving AI law and policy? And how do you think your work at OpenAI contributes to those goals?

Comment by JoshYou on [Link] A new charity evaluator (NYTimes) · 2019-11-27T06:11:04.596Z · EA · GW

Seems like its mission sits somewhere between GiveWell's and Charity Navigator's. GiveWell studies a few charities to find the very highest impact ones according to its criteria. Charity Navigator attempts to rate every charity, but does so purely on procedural considerations like overhead. ImpactMatters is much broader and shallower than GiveWell but unlike Charity Navigator does try to tell you what actually happens as the result of your donation.

Comment by JoshYou on Has any EA oriented organization tried promoting donors on their social media? · 2019-10-29T16:06:49.664Z · EA · GW

I think I would be more likely to share my donations this way compared to sharing them myself, because it would feel easier and less braggadocious (I currently do not really advertise my donations).

Comment by JoshYou on How do you, personally, experience "EA motivation"? · 2019-08-18T02:03:47.065Z · EA · GW

Among other things, I feel a sense of pride and accomplishment when I do good, the way I imagine that someone who cares about, say, the size of their house feels when they think about how big their house is.

Comment by JoshYou on Four practices where EAs ought to course-correct · 2019-08-03T22:55:19.760Z · EA · GW

Absolutely, EAs shouldn't be toxic, inaccurate, or uncharitable on Twitter or anywhere else. But I've seen a few examples of people effectively communicating about EA issues on Twitter, such as Julia Galef and Kelsey Piper, at a level of fidelity and niceness far above the average for that website. On the other hand they are briefer, more flippant, and spend more time responding to critics outside the community than they would on other platforms.

Comment by JoshYou on Four practices where EAs ought to course-correct · 2019-07-30T23:36:10.564Z · EA · GW

Yep, though I think it takes a while to learn how to tweet, whom to follow, and whom to tweet at before you can get a consistently good experience on Twitter and avoid the nastiness and misunderstandings it's infamous for.

There's a bit of an extended universe of Vox writers, economists, and "neoliberals" that are interested in EA and sometimes tweet about it, and I think it would be potentially valuable to add some people who are more knowledgeable about EA into the mix.

Comment by JoshYou on Four practices where EAs ought to course-correct · 2019-07-30T22:50:50.638Z · EA · GW

On point 4, I wonder if more EAs should use Twitter. There are certainly many options to do more "ruthless" communication there, and it might be a good way to spread and popularize ideas. In any case it's a pretty concrete example of where fidelity vs. popularity and niceness vs. aggressive promotion trade off.

Comment by JoshYou on What Do Unconscious Processes in Humans Tell Us About Sentience? · 2019-06-15T15:52:22.070Z · EA · GW

This all seems to assume that there is only one "observer" in the human mind, so that if you don't feel or perceive a process, then that process is not felt or perceived by anyone. Have you ruled out the possibility of sentient subroutines within human minds?

Comment by JoshYou on Is preventing child abuse a plausible Cause X? · 2019-05-06T05:09:23.680Z · EA · GW

Sadly, Jiwoon passed away last year.

Comment by JoshYou on [Question] Pros/Cons of Donor-Advised Fund · 2019-04-22T21:16:27.483Z · EA · GW

Some links if you haven't seen them yet:

I don't use a DAF but I've considered it in the past. In my view, the chief advantage is that they allow you to claim the tax deduction when you deposit money into the DAF, before you actually make the donation. They're also exempt from capital gains taxes, although you can also avoid capital gains taxes by donating appreciated assets directly to the charity, but that depends on whether the organization will accept them (not sure how universal this is). They also charge fees, which can be fairly expensive but are cheaper than capital gains taxes on expectation.

Comment by JoshYou on Should EA grantmaking be subject to independent audit? · 2019-04-18T02:47:20.761Z · EA · GW

Open Phil would be a good candidate for this, though that's a difficult proposition due to its sheer size. It is a somewhat odd situation that Open Phil throws huge amounts of money around, much of which happens without any comment from the EA community.

Comment by JoshYou on Why is the EA Hotel having trouble fundraising? · 2019-03-26T23:42:56.798Z · EA · GW

I wonder if the lack of tax deductibility and the non-conventional fundraising platform (GoFundMe) nudge people into not donating or donating less than they would to a more respectable-seeming charity.

(As a tangent, there's a donation swap opportunity for the EA Hotel that most people are probably not aware of).

Comment by JoshYou on EA Hotel Fundraiser 3: Estimating the relative Expected Value of the EA Hotel (Part 1) · 2019-03-13T00:07:55.997Z · EA · GW

Speaking as someone with a undergrad degree in math, I would have found a non-technical summary for this post to be helpful. So I expect this would apply much more to many other forum readers.

Comment by JoshYou on After one year of applying for EA jobs: It is really, really hard to get hired by an EA organisation · 2019-02-28T20:14:10.246Z · EA · GW

For one of the work tests I did for Open Phil, the instruction sheet specifically asked that the work test not be shared with anyone. That might have been intended as a temporary restriction, I'm not sure, but I'm not planning on sharing it unless I hear otherwise.

Comment by JoshYou on Vox's "Future Perfect" column frequently has flawed journalism · 2019-01-26T15:23:35.448Z · EA · GW

Agreed. I don't see any "poor journalism" in any of the pieces mentioned. A few of them would be "poor intervention reports" if we chose to judge them by that standard.

Comment by JoshYou on Climate Change Is, In General, Not An Existential Risk · 2019-01-12T03:30:54.309Z · EA · GW

It's clear that climate change has at best a small probability (well under 10%) of causing human extinction, but many proponents of working on other x-risks like nuclear war and AI safety would probably give low probabilities of human extinction for those risks as well. I think the positive feedback scenarios you mention (permafrost, wetlands, and ocean hydrates) deserve some attention from an x-risk perspective because they seem to be poorly understood, so the upper bound on how severe they might be may be very high. You cite one simulation that burning all available fossil fuels would increase temperatures by 10 °C, but that isn't necessarily an upper bound because there are non-fossil fuel sources carbon on Earth that could be released to the atmosphere. It would of course also be necessary to estimate how high the extinction risk conditional on various levels of extreme warming (8°C, 10°C, 15°C, 20°C?) would be.

Regardless, it's a good idea to have a clear view of how big the risk is. You're right that the casual claims about extinction or planetary uninhabitability I hear from many people who are concerned about climate change are not justified, and they seem a bit irresponsible.

Comment by JoshYou on How should large donors coordinate with small donors? · 2019-01-10T03:48:24.082Z · EA · GW

Holden also wrote (by the way, I think your link is broken):

We fully funded things we thought were much better than the "last dollar" (including certain top charities grants) but not things we thought were relatively close when they also posed coordination issues. For this case, fully funding top charities would have had pros and cons relative to splitting: we think the dollars we spent would've done slightly more good, but the dollars spent by others would've done less good (and we think we have a good sense of the counterfactual for most of those dollars). We guessed that the latter outweighed the former.

So an important crux here is the proportion of small-donor money to e.g. GiveWell charities that would be crowded out into much less effective charities or to new projects with high expected value. For reference, GiveWell has moved about $30-40 million a year in small donations. I am not sure what proportion of that comes from people who are not closely aligned/affiliated with the EA community, but I would guess it's the majority.

I would question whether Holden is correct though. Global health/development is a big space, so if Good Ventures increased funding to GiveWell top charities by a lot, GiveWell would still exist and would move their recommendations over to interventions that aren't fully funded yet. For example, cash transfers seemingly could absorb a lot of money, and the Gates Foundation probably moves more to global poverty causes every year than GoodVentures will spend per year at its peak. The claim seems to depend on small GiveWell donors being excited by GiveWell's specific top charities right now, such that they would not give to GiveWell top charities if the current top charities were fully funded and GiveWell issued new recommendations, and would instead give to charities even less effective than these new top charities. That might be true if donors are really motivated by the headline cost-per-life-saved number rather than being attracted by GiveWell's research and methodology. I don't have a very strong intuition either way, so I'd be curious if someone more knowledgeable could shed some light.

Comment by JoshYou on EA orgs are trying to fundraise ~$10m - $16m · 2019-01-06T14:56:09.321Z · EA · GW

If we're using these numbers to inform whether EA is funding constrained, it would be good if someone followed up and figured out how much these organizations actually ended up raising.

Comment by JoshYou on Challenges in Scaling EA Organizations · 2018-12-21T23:32:41.091Z · EA · GW

One thing I've wondered about is what the optimal rate at which new EA organizations should be founded, and whether that's an effective way around growth bottlenecks. For example, Rethink Priorities has grown rapidly this year, and it doesn't seem likely that that growth would have happened anyway within previously existing organizations had Rethink Priorities not been founded.

Comment by JoshYou on Animal Welfare Fund AMA · 2018-12-20T01:29:09.665Z · EA · GW

This fund has seemingly taken a very "hits-based" approach to funding small, international grassroots organizations. How do you plan on evaluating and learning from these grants?

Comment by JoshYou on Long-Term Future Fund AMA · 2018-12-20T01:14:56.644Z · EA · GW

This post contains an extensive discussion on the difficulty of evaluating AI charities because they do not share all of their work due to info hazards (in the "Openness" section as well as the MIRI review). Will you have access to work that is not shared with the general public, and how will you approach evaluating research that is not shared with you or not shared with the public?

Comment by JoshYou on Long-Term Future Fund AMA · 2018-12-20T01:11:00.715Z · EA · GW

Under what conditions would you consider making a grant directed towards catastrophic risks other than artificial intelligence?

Comment by JoshYou on New web app for calibration training funded by the Open Philanthropy Project · 2018-12-17T05:16:31.475Z · EA · GW

Vague or context-less questions might help you calibrate your views on topics you know very little about?

I am now somewhat better calibrated at claims about European football than I was before, I guess.

Comment by JoshYou on 2017 Donor Lottery Report · 2018-11-16T00:30:52.661Z · EA · GW

This is a great writeup and also a good demonstration of the value of donor lotteries. Is CEA planning on running another one anytime soon? Their lotteries page just says "There are currently no active lotteries". I think the lottery experiments have gone well and this should be a regular thing, unless running a lottery consumes a lot of staff time or has some other large cost.

Comment by JoshYou on One for the World as a potential vehicle to expand the reach of Effective Altruism · 2018-08-02T02:10:56.397Z · EA · GW

On the flip side, maybe it's a good idea for 1FTW to maintain some distance from the EA community/EA as a concept. If they specialize in promoting effective giving to global poverty to people who are unlikely to embrace EA as a whole, that might be a good way to avoid competing with existing EA outreach.

Comment by JoshYou on The EA Community and Long-Term Future Funds Lack Transparency and Accountability · 2018-08-01T23:11:47.647Z · EA · GW

It seems that Nick has not been able to leverage his position as EA fund manager to outperform his Open Phil grants (or at least meaningfully distinguish his EA fund grants from his Open Phil grants). This means that we can think of donating to the far future and community funds as having similar cost-effectiveness to individual donations to Open Phil earmarked for those causes. This seems like a problem, since the best individual donations should be able to outperform Open Phil, at least when you account for the benefits of not centralizing donations on too few decisionmakers. I don't see anyone calling for Open Phil to accept/solicit money from small donors.

The case for finding another manager seems pretty strong. EA funds is a fundamentally sound idea - we should be trying to consolidate donation decisions somewhat to take advantage of different levels of expertise and save small donors' time and mental energy, but this doesn't seem like the best way to do it.

Comment by JoshYou on The EA Community and Long-Term Future Funds Lack Transparency and Accountability · 2018-07-23T01:35:40.967Z · EA · GW

Lewis announced another round of grants for the Animal Welfare fund on Facebook on June 26, though it's not clear when exactly the grants were paid out or will be paid out. The Animal Welfare fund page has not been updated with this information. This seems surprising since Lewis has already written up an explanation of the grant; it just isn't on the website yet.

Comment by JoshYou on Announcing the Effective Altruism Handbook, 2nd edition · 2018-05-03T04:33:37.800Z · EA · GW

Doing Good Better is more accessible and spends a lot more time introducing and defending the basic idea of EA instead of branching out into more advanced ideas. It is also much more focused on global poverty.

Comment by JoshYou on How to improve EA Funds · 2018-05-01T05:00:39.367Z · EA · GW

I just noticed that Nick posted updates on the Community Fund and Far Future Fund pages (it's the same update on both pages) on 4/24. I'm commenting here for visibility since I have not seen these updates advertised anywhere.

Comment by JoshYou on 80,000 Hours: EA and Highly Political Causes · 2017-01-29T16:27:43.867Z · EA · GW

Support for a cause area isn't bias. That's just having an opinion. Your argument would imply that ACE is biased because they are run by animal activists, or that Givewell is biased because they advocate for reducing global poverty. These groups aren't necessarily an authority when you're deciding between cause areas, of course. But in deciding which organization is most effective within a given cause area, the "trusted experts" are almost always going to be advocates for that cause area.

More generally, you keep trying to frame your points as politically neutral "meta" considerations but it definitely feels like you have an axe to grind against the activist left which motivates a lot of what you're saying.

Comment by JoshYou on Charity Science Effective Legacies · 2016-12-30T18:34:19.483Z · EA · GW

The title seems a little... harsh.

Comment by JoshYou on We Must Reassess What Makes a Charity Effective · 2016-12-24T15:59:53.672Z · EA · GW

These are pretty unoriginal generic arguments against developing-world charity. I think you should do more research on how these arguments apply GiveWell charities and engage with the existing arguments they have made for why their charities are cost-effective. Local mosquito net industries are clearly not an important driver of economic growth that they would outweigh the benefit of large reductions in malaria. The second point is just a quote about a bad charity methodology with almost no explanation for why GiveWell charities do what Easterly criticizes. The third point is just wrong. GiveDirectly gives one-time cash transfers to individuals, not ongoing aid.

Comment by JoshYou on Contra the Giving What We Can pledge · 2016-12-05T02:25:48.763Z · EA · GW

I'm still pretty confused about why you think donating 10% has to be time-confusing. People who outsource their donation decisions to, say, Givewell might only spend a few hours a year (or a few minutes, depending on how literally we interpret "outsourcing) deciding where to donate.

Comment by JoshYou on Ethical offsetting is antithetical to EA · 2016-01-06T02:50:11.552Z · EA · GW

"And as Scott Alexander points out, offsetting could lead people to think it’s acceptable to do big harmful things as long as they offset them."

I think it would be helpful to distinguish between the claims (1) "given that one has imposed some harm, one is obligated to offset it" and (2) "any imposition of harm is justified if it is offset." This article argues against the first claim, while Scott argues that the second one seems false. It seems pretty easy to imagine someone accepting (1) and rejecting (2), and I'd be pretty skeptical of a causal connection between promoting (1) and more people believing in (2). The reverse seems just as (un)likely: "hey, if I don't have to offset my harms, maybe causing harm doesn't really matter to begin with."

Comment by JoshYou on Ideas for new experimental EA projects you could fund! · 2014-12-03T00:40:54.026Z · EA · GW

Hire a full or part-time Personal Assistant for Prof Nick Bostrom

Is there a reason this couldn't be done with FHI funding? If FHI believed that this was the best use of an additional [however much it takes to hire an assistant], then an unrestricted donation of that amount would make it happen. If not, it's much less clear that this would be a good idea.

Comment by JoshYou on The new GiveWell recommendations are out: here's a summary of the charities · 2014-12-01T18:41:55.294Z · EA · GW

DtWI has a relatively small funding gap of $1.3 million.

Comment by JoshYou on Introduce Yourself · 2014-09-18T18:41:12.125Z · EA · GW

Hi, I'm Josh. I'm a sophomore in college and I'm tentatively planning on EtG through programming. I have been donating to CEA for movement-building purposes, but may switch to ACE and/or ACE-recommended charities in the near future. I became an EA after being heavily exposed to moral philosophy (esp. utilitarianism) through doing debate in high school.

When I'm not doing school work I enjoy playing video games, reading philosophy, working out, and programming.

Comment by JoshYou on Open Thread · 2014-09-16T02:26:04.135Z · EA · GW

William Macaskill makes a few good points here about why EA does not rely on utilitarianism. It's true that a lot of EAs are utilitarian, but I've seen plenty of discussions on normative ethics among EA circles, so I wouldn't describe it as a silent unanimity.