[Job] Policy and Advocacy Coordinator at Technologies for Pandemic Defense 2023-02-23T00:39:02.626Z
A hypothesis for why some people mistake EA for a cult 2022-05-12T18:39:04.090Z
A visualization of some orgs in the AI Safety Pipeline 2022-04-10T16:52:44.169Z
What are some heuristics for longtermist project evaluation? 2022-02-07T05:55:52.278Z
[Creative Writing Contest] The Rise of The Effective Shoppers 2021-10-21T21:37:35.349Z
[Creative Writing Contest] The Legend of the Goldseeker 2021-10-21T21:31:23.105Z
Introducing a project on accountability in governance, plus a call for volunteers 2020-12-30T17:55:11.551Z


Comment by Aman Patel (aman-patel) on Announcing Nonlinear Emergency Funding · 2022-11-14T15:58:36.494Z · EA · GW

I'm curious about what's the original source of the funding you're giving out here. According to this Nonlinear received $250k from Future Fund and $600k from Survival and Flourishing Fund. Is the funding being distributed here coming solely from the SFF grant? Does Nonlinear have other funding sources besides Future Fund and SFF? 

(I didn't do any deeper dive than looking at Nonlinear's website, where I couldn't find anything about funding sources.)

Comment by Aman Patel (aman-patel) on My Most Likely Reason to Die Young is AI X-Risk · 2022-07-04T20:19:54.121Z · EA · GW

Thanks for writing this--even though I've been familiar with AI x-risk for a while, it didn't really hit me on an emotional level that dying from misaligned AI would happen to me too, and not just "humanity" in the abstract. This post changed that. 

Might eventually be useful to have one of these that accounts for biorisk too, although biorisk "timelines" aren't as straightforward as trying to estimate the date that humanity builds the first AGI.

Comment by aman-patel on [deleted post] 2022-05-30T00:14:03.630Z

Thanks for posting your attempt! Yeah, it does seem like you ran into some of those issues in your attempt, and it's useful information to know that this task is very hard. I guess one lesson here is that we probably won't be able to build perfect institutions on the first try, even in safety-critical cases like AGI governance.

Comment by aman-patel on [deleted post] 2022-05-26T20:53:56.814Z

Just stumbled upon this post--I like the general vein in which you're thinking. Not sure if you're aware of it already, but this post by Paul Christiano addresses the "inevitable dangerous technology" argument as it relates to AI alignment. 

 - "First-principles design is intractable and misses important situation-specific details" - This could easily be true, I don't have a strong opinion on it, just intutions.

I think this objection is pretty compelling. The specific tools that an institution can use to ensure that a technology is deployed safely will ultimately depend on the nature of that technology itself, its accessibility/difficulty of replication, the political/economic systems it's integrated into, and the incentives surrounding its deployment. (Not an exhaustive list.)

Usually, any type of regulation or "responsible power-wielding" comes with tradeoffs (to freedom, efficiency, equitability, etc.), and it'll be hard to assess whether these accepting these tradeoffs is prudent without a specific technology in mind.

That said, I think it can still be a worthwhile exercise to think about how we can build governance practices that are robust to worst-case scenarios for all of the above. I can imagine some useful insights coming out of that kind of exercise!

Comment by Aman Patel (aman-patel) on EA culture is special; we should proceed with intentionality · 2022-05-22T18:09:23.587Z · EA · GW

Thanks, great points (and counterpoints)!

If you are a community builder (especially one with a lot of social status), be loudly transparent with what you are building your corner of the movement into and what tradeoffs you are/aren’t willing to make.

I like this suggestion--what do you imagine this transparency looks like? Do you think, e.g., EA groups should have pages outlining their community-building philosophies on their websites? Should university groups should write public Forum posts about their plans and reasoning before every semester/quarter or academic year? Would you advocate for more community-building roundtables at EAGs? (These are just a few possible example modalities of transparency that just came to my head, very interested in hearing more.)

Comment by Aman Patel (aman-patel) on A hypothesis for why some people mistake EA for a cult · 2022-05-21T22:17:25.373Z · EA · GW

Yeah, I've had several (non-exchange) students ask me what altruism means--my go-to answer is "selflessly helping others," which I hope makes it clear that it describes a practice rather than a dogma. 

Comment by Aman Patel (aman-patel) on A hypothesis for why some people mistake EA for a cult · 2022-05-21T22:12:03.860Z · EA · GW

Thanks for the comment! I agree with your points--there are definitely elements of EA, whether they're core to EA or just cultural norms within the community, that bear stronger resemblances to cult characteristics. 

My main point in this post was to explore why someone who hasn't interacted with EA before (and might not be aware of most of the things you mentioned) might still get a cult impression. I didn't mean to claim that the Google search results for "altruism" are the most common reason why people come away with a cult impression. Rather, I think that they might explain a few perplexing cases of cult impressions that occur before people become more familiar with EA. I should have made this distinction clearer, thanks for pointing it out :)

Comment by Aman Patel (aman-patel) on Which Post Idea Is Most Effective? · 2022-04-26T22:02:39.731Z · EA · GW

Hey Jordan! Great to see another USC person here. The best writing advice I've gotten (that I have yet to implement) is to identify a theory of change for each potential piece--something to keep in mind!

6 sounds interesting, if you can make a strong case for it. Aligning humans isn't an easy task (as most parents, employers, governments, and activists know very well), so I'm curious to hear if you have tractable proposals.

7 sounds important given that a decent number of EAs are vegan, and I'm quite surprised I haven't heard of this before. 15 IQ points is a whole standard deviation, so I'd love to see the evidence for that.

8 might be interesting. I suspect most people are already aware of groupthink, but it could be good to be aware of other relevant phenomena that might not be as widely-known (if there are any).

From what I can tell, 11 proposes a somewhat major reconsideration of how we should approach improving the long-term future. If you have a good argument, I'm always in favor of more people challenging the EA community's current approach. I'm interested in 21 for the same reason.

(In my experience, the answer to 19 is no, probably because there isn't a clear, easy-to-calculate metric to use for longtermist projects in the way that GiveWell uses cost-effectiveness estimates.)

Out of all of these, I think you could whip up a draft post for 7 pretty quickly, and I'd be interested to read it!

Comment by Aman Patel (aman-patel) on What are some heuristics for longtermist project evaluation? · 2022-02-11T00:35:42.411Z · EA · GW

Thanks Linch! This list is really helpful. One clarifying question on this point: 

Relatedly, what does the learning/exploration value of this project look like?

  1. To the researcher/entrepreneur?
  2. To the institution? (if they're working in an EA-institutional context)
  3. To the EA or longtermist ecosystem as a whole?

For 1) and 2), I assume you're referring to the skills gained by the person/institution completing the project, which they could then apply to future projects. 

For 3), are you referring to the possibility of "ruling out intervention X as a feasible way to tackle x-risks"? That's what I'm assuming, but I'm just asking to make sure I understand properly.

Thanks again!

Comment by Aman Patel (aman-patel) on High School Seniors React to 80k Advice · 2021-12-20T15:07:44.351Z · EA · GW

This thinking has come up in a few separate intro fellowship cohorts I’ve facilitated. Usually, somebody tries to flesh it out by asking whether it’s “more effective” to save one doctor (who could then be expected to save five more lives) or two mechanics (who wouldn’t save any other lives) in trolley-problem scenarios. This discussion often gets muddled, and many people have the impression that “EAs” would think it’s better to save the doctor, even though I doubt that’s a consensus opinion among EAs. I’ve found this to be a surprisingly large snag point that isn’t discussed much in community-building circles.

I think it would be worth it to clarify the difference between intrinsic and instrumental value in career advice/intro fellowships/other first interactions with the EA community, because there are some people who might agree with other EA ideas but find that this argument undermines our basic principles (as well as the claim that you don’t need to be utilitarian to be an EA). Maybe we could extend current messaging about ideological diversity within EA.

That said, I read Objection 4 differently. Many people (especially in cultures that glorify work) tie their sense of self-worth to their jobs. I don’t know how universal this is, but at least in my middle-class American upbringing, there was a strong sense that your career choice and achievement is a large part of your value as a person. 

As a result, some people feel personally judged when their intended careers aren’t branded as “effective”. If you equate your career value with your personal value, you won’t feel very good if someone tells you that your career isn’t very valuable, and so you’ll resist that judgment.

I don’t think that this feeling precludes people from being EAs. It takes time to separate yourself from your current or intended career, and Objection 4 strikes me as a knee-jerk defensive reaction. Students planning to work in shipping logistics won’t immediately like the idea that the job they’ve been working hard to prepare for is “ineffective,” but they might come around to it after some deeper reflection. 

I could be misreading Objection 4, though. It could also mean something like “shipping logistics is valuable because the world would grind to a halt if nobody worked in shipping logistics,” but then that’s just a variant of Objection 5.

I’m very curious to know more about the sense in which these students gave Objection 4. 

Comment by Aman Patel (aman-patel) on [Creative Writing Contest] The Legend of the Goldseeker · 2021-11-09T16:13:03.685Z · EA · GW

Changed "guilt" to "responsibility," but I'm not sure if that's much better.

Comment by Aman Patel (aman-patel) on [Creative Writing Contest] The Legend of the Goldseeker · 2021-11-09T16:11:45.026Z · EA · GW

Thanks for the feedback! I think this is probably a failure of the story more than a failure of your understanding--after all, a story that's hard to understand isn't fulfilling its purpose very well. Jackson Wagner's comment below is a good summary of the main points I was intending to get across.

Next time I write, I'll try to be more clear about the points I'm trying to convey. 

Comment by Aman Patel (aman-patel) on [Creative Writing Contest] The Legend of the Goldseeker · 2021-11-09T15:59:23.416Z · EA · GW

"As tagged, this story strikes me as a fable intended to explain one of the mechanisms behind so-called "S-risks", hellish scenarios that might be a fate worse than the "death" represented by X-risks."

That's what I was going for, although I'm aware that I didn't make this as clear as I should have.

"Of course it's a little confusing to have the twist with the sentient birds -- I think rather than a literal "farmed animal welfare" thing, this is intended to showcase a situation where two different civilizations have very different values."

Same thing here. This is what I was trying to get at, but couldn't think of many other scenarios involving suffering agents where one group of people cares and another doesn't.

"I don't really understand why the story is a frame story, or why the main purpose of the ritual is for all the Kunus to feel "collective guilt"... EA is usually trying to steer away from giving the impression that we want everyone to feel guilty all the time."

This is really helpful feedback--I didn't realize that "collective guilt" came across as the point of the story, and I definitely agree that making people feel guilty is counterproductive. I can't remember why I threw in that phrase (probably because I couldn't think of anything else), but I'll change it now. 

Totally unrelated point, but I thought the economics of this story were a little wacky.

Yup, definitely more than a "little" wacky :) Maybe using another resource like food or water or land would be better--but then it would have been harder to make the point that each country thought were doing the right thing.

This is a good part of the parable -- if S-risks ever occur, the civilizations that commit those galactic war crimes will probably be convinced of their righteousness, and indeed probably won't even recognize that they are committing a wrong.

This is the central point that I wanted to get across. Whether we're considering a civilization or an advanced AI, s-risks need not result from intentional malevolence. I'm glad it didn't get too distorted, but it seems like there are better ways to build a story around this point.

Another side-note: a lot of the ideas behind this story are discussed in the Center on Long-Term Risk's research agenda. I don't know whether they would agree with my presentation or conceptualization of those ideas.

Thank you so much for the feedback!

Comment by Aman Patel (aman-patel) on [Creative Writing Contest] The Rise of The Effective Shoppers · 2021-10-24T02:23:46.010Z · EA · GW


Comment by Aman Patel (aman-patel) on [Creative Writing Contest] The Rise of The Effective Shoppers · 2021-10-24T02:23:11.142Z · EA · GW

Thanks! I'm glad you enjoyed it. The main reason I wrote this was to practice creative writing--and the Forum contest seemed to be a good place to do that. This is the first time I tried writing short stories--the only other creative writing piece I've published anywhere is this one, which I also wrote for the Forum contest:

I hope that helps!

Comment by Aman Patel (aman-patel) on How to Train Better EAs? · 2021-10-05T23:09:08.764Z · EA · GW

I recently learned about Training for Good, a Charity Entrepreneurship-incubated project, which seems to address some of these problems. They might be worth checking out.

I think this is a great exercise to think about, especially in light of somewhat-recent discussion on how competitive jobs at EA orgs are. There seems to be plenty of room for more people working on EA projects, and I agree that it’s probably good to fill that opportunity. Some loose thoughts:

There seem to be two basic ways of getting skilled people working on EA cause areas:
1.  Selectively recruiting people who already have skills.
2. Recruiting promising people who might not yet have needed skills and train them. 

Individual organizations can choose both options, depending on their level of resources. But if most organizations choose option 1, the EA community might be underutilizing its potential pool of human resources. So we might want the community in general to use option 2, so that everyone who wants to be involved with EA can have a role—even if individual EA organizations still choose option 1. For this to happen, the EA community would probably need a program whereby motivated people can choose a skillset to learn, are taught that skillset, and are matched with a job at the end of the process. 

Currently, motivated people who don’t yet possess skills are placed into a jumble of 1-on-1 conversations, 80k advising calls, and fellowship and internship listings. Having those calls and filling out internship and fellowship applications takes a ton of time and mental energy, and might leave people more confused than they were initially. A well-run training program could eliminate many of these inefficiencies and reduce the risk that interested people won’t be able to find a job in EA. 

We can roughly rank skill-building methods by the number of people they reach (“scale”), and the depth of training that they provide. In the list below, “high depth” skill development could lead to being hired for that skill (when one would not have been hired for that skill otherwise), “medium depth” as warranting a promotion or increase in seniority level, and “low depth” as an enhancement of knowledge that can help someone perform their job better, but probably won’t lead to new positions or higher status.

  • Internal development within organizations, like Aaron Gertler mentioned (small scale, medium depth)
  • Internship/fellowship programs (medium scale, medium depth)
  • One-off workshops and lectures (small scale, low depth)
  • Cause area-specific fellowships, like EA Cambridge's AGI Safety Fellowship (large scale, low depth) 
  • A training program like the one I described above (large scale, high depth)
  • An EA university, as proposed here (large scale, high depth)

If we choose option 2, we probably want large scale, high depth ways to train people. I’m interested in hearing people’s thoughts on whether this is a good way to evaluate skill-building methods.

One caveat: there’s a lot more interest in working for the military than there is in working for EA orgs. Since this interest already exists, the military just needs to capitalize on it (although they still spend lots of money on recruitment ads and programs like ROTC). The EA community doesn’t even have great name recognition, so it’s probably premature to assume that we’d have waves of people signing up for such a training program—but it’s possible that we could get to that point with time.

Comment by Aman Patel (aman-patel) on What we learned from a year incubating longtermist entrepreneurship · 2021-09-26T18:31:34.065Z · EA · GW

Thanks for this post! Reading through these lessons has been really informative. I have a few more questions that I'd love to hear your thinking on:

1) Why did you choose to run the fellowship as a part-time rather than full-time program?

2) Are there any particular reasons why fellowship participants tended to pursue non-venture projects?

3) Throughout your efforts, were you optimizing for project success or project volume, or were you instead focused on gathering data on the incubator space?

4) Do you consider the longtermist incubation space to be distinct from the x-risk reduction incubation space?

5) Was there a reason you didn't have a public online presence, or was it just not a priority?

Comment by Aman Patel (aman-patel) on Should Chronic Pain be a cause area? · 2021-05-26T19:45:02.831Z · EA · GW

Thanks for the post, this is an important and under-researched topic. 

Examples include some well-known conditions (chronic migraine, fibromyalgia, non-specific low-back pain), as well as many lesser-known ones (trigeminal neuralgia, cluster headache, complex regionary pain syndrome)

Some of these well-known chronic pain conditions can be hard to diagnose, too. Chronic pain conditions like fibromyalgia, ME/CFS, rheumatoid arthritis, and irritable bowel syndrome are frequently comorbid with each other, and may also be related to depression and mental health disorders. This overlap probably makes it harder for doctors to tease out the root cause of patients’ symptoms.

As an anecdote, a close relative spent around a year bouncing around various doctors before she got a useful diagnosis, and even then the recommended therapies didn’t help much. So far, her pain is managed best by a diet she found on the internet herself.

I speculate that conventional medicine’s relative lack of machinery for identifying and treating some of these chronic illnesses may cause some patients to turn to pseudoscience instead—which could be another downstream harm of neglecting chronic pain treatments. (I haven’t tried to look for evidence for/against this conclusion.) 

Comment by Aman Patel (aman-patel) on saulius's Shortform · 2021-05-25T04:20:30.056Z · EA · GW

This is an interesting idea. I'm trying to think of it in terms of analogues: you could feasibly replace "digital minds" with "animals" and achieve a somewhat similar conclusion. It doesn't seem that hard to create vast amounts of animal suffering (the animal agriculture industry has this figured out quite well), so some agent could feasibly threaten all vegans with large-scale animal suffering. And as you say, occasionally following through might help make that threat more credible. 

Perhaps the reason we don't see this happening is that nobody really wants to influence vegans alone. There aren't many strategic reasons to target an unorganized group of people whose sole common characteristic is that they care about animals. There isn't much that an agent could gain from a threat.

I imagine the same might be true of digital minds. If it's anything similar to the animal case, moral circle expansion to digital minds will likely occur in the same haphazard, unorganized way--and so there wouldn't be much of a reason to specifically target people who care about digital minds. That said, if this moral circle expansion caught on predominantly in one country (or maybe within one powerful company), a competitor or opponent might then have a real use for threatening the digital mind-welfarists. Such an unequal distribution of digital mind-welfarists seems quite unlikely, though.

At any rate, this might be a relevant consideration for other types of moral circle expansion, too.

Comment by Aman Patel (aman-patel) on Introducing a project on accountability in governance, plus a call for volunteers · 2021-01-12T19:55:16.354Z · EA · GW

Thanks for the tip! I'll try contacting him through the website you linked--it would be great to hear more from people who have attempted this sort of project before.

Comment by Aman Patel (aman-patel) on AMA: Owen Cotton-Barratt, RSP Director · 2020-08-29T16:31:17.207Z · EA · GW

How do you think the EA community can improve its interactions and cooperation with the broader global community, especially those who might not be completely comfortable with the underlying philosophy? Do you think it's more of a priority to spread those underlying arguments, or to simply grow the network of people sympathetic to EA causes, even if they disagree with the principles of EA?

Comment by Aman Patel (aman-patel) on Open and Welcome Thread: August 2020 · 2020-08-29T16:17:18.258Z · EA · GW

Hi everyone! I'm Aman, an undergrad at USC currently majoring in computational neuroscience (though that might change). I'm very new to EA, so I haven't yet had the chance to be involved with any EA groups, but I would love to start participating more with the community. I found EA after spending a few months digging into artificial general intelligence, and it's been great to read everyone's thoughts about how to turn vague moral intuitions into concrete action plans.

I have a soft spot for the standard big-picture philosophy/physics topics, like the nature of intelligence and meta-ethics and epistemology and theories of everything, but also the wildly unpragmatic questions (like whether we might consider directing ourselves into a time loop once heat death comes around, if it's possible).

As a career, I tentatively want to focus on improving global governance capacity, since I'm inclined to think that it might ultimately determine how well EA-related research and prioritization can be implemented (and also how well we are able to handle x- and s-risks, and capitalize on safe AI). I realize that this is probably one of the least tractable goals to have, so I might end up working in another area, like international development, mental health, science policy, or something else entirely. Amusingly, all the EA career advice out there has only made me more confused about what I should be doing (but I'm probably approaching it wrong).

Anyway, I'm excited to be here and grateful for the opportunity to start interacting with the EA community!