edoarad's Shortform 2019-08-16T13:35:05.296Z · score: 2 (1 votes)
Microsoft invests 1b$ in OpenAI 2019-07-22T18:29:57.316Z · score: 21 (9 votes)
Cochrane: a quick and dirty summary 2019-07-14T17:46:42.945Z · score: 11 (7 votes)
Target Malaria begins a first experiment on the release of sterile mosquitoes in Africa 2019-07-05T04:58:44.912Z · score: 8 (5 votes)
Babbling on Singleton Governance 2019-06-23T04:59:30.567Z · score: 1 (2 votes)
Is there an analysis that estimates possible timelines for arrival of easy-to-create pathogens? 2019-06-14T20:41:42.228Z · score: 12 (5 votes)
Innovating Institutions: Robin Hanson Arguing for Conducting Field Trials on New Institutions 2019-03-31T20:33:06.581Z · score: 8 (4 votes)
China's Z-Machine, a test facility for nuclear weapons 2018-12-13T07:03:22.910Z · score: 12 (6 votes)


Comment by edoarad on edoarad's Shortform · 2019-08-16T13:35:05.426Z · score: 2 (2 votes) · EA · GW

Summary: academia has a lot of problems and it could work much better. However, these problems are not as catastrophic as an outside perspective would suggest. My (contrarian, I guess) intuition is that scientific progress in biology is not slowing down. Specific parts of academia that seem to be problematic: rigid, punishing for deviation, career progression; peer review; need to constantly fundraise for professors. Parts that seem to be less of a problem than I initially thought: short-termism; lack of funding for young scientists.

Comment by edoarad on What types of organizations would be ideal to be distributing funding for EA? (Fellowships, Organizations, etc) · 2019-08-05T03:52:28.196Z · score: 3 (3 votes) · EA · GW

For me, a research organization that can set a global research agenda and direct the research, foster collaborations and give funding for the type of research that is underfunded in academia will be very helpful.

I see two different goals. One is coordination of researchers. The other is to have a safety net that allows to reduce the academic incentives to publish and to be more of a specialist.

Comment by edoarad on The Possibility of an Ongoing Moral Catastrophe (Summary) · 2019-08-04T04:42:24.836Z · score: 3 (3 votes) · EA · GW

Thanks for the summary. I have two takeaways:

1. EA is (in part) claiming that there are several ongoing moral catastrophes caused by inaction against global poverty, animal suffering, x-risk,... (some of them are definitely caused by action, but that does not matter as much on consequentialist grounds). Unknown ongoing moral catastrophes are cause-X.

2. The possibility of working to increase our capability to handle undiscovered ongoing moral catastrophe in the future as a major goal. The idea I saw here was to reserve resources, which is a very interesting argument to invest in economic growth.

Comment by edoarad on Experiments in GiveWell communication · 2019-07-30T17:40:13.875Z · score: 1 (1 votes) · EA · GW

From the onset I was expecting that in the last section you would ask for help from EAs with experience in design or advertising or something on that route. I guess that you have capable people on that front. Figured it is interesting to note that, regarding career paths in design.

And thank you for your amazing work :)

Comment by edoarad on William Rathbone: 19th-century effective altruist? · 2019-07-30T16:55:36.922Z · score: 2 (2 votes) · EA · GW

What is known about how these ideas were recieved?

Comment by edoarad on Latest EA Updates for July 2019 · 2019-07-27T07:56:50.684Z · score: 6 (5 votes) · EA · GW

For anyone interested, the Horizon Europe Survey took me about 40 minutes to complete but could have been faster. It seems like a place to raise important EA-aligned topics.

Not sure how impactful this is, but I assume that there won't be too many replies. Anyway it was a chance for me to understand the Horizon Europe program better and my attitude toward many relevant topics.

Comment by edoarad on Ought: why it matters and ways to help · 2019-07-26T08:47:51.432Z · score: 3 (2 votes) · EA · GW

The "Updates" page is actually at (but is displayed as ).

Comment by edoarad on Cochrane: a quick and dirty summary · 2019-07-14T17:49:51.939Z · score: -3 (4 votes) · EA · GW

So that took 18 minutes... Spent a lot of time on making sure I'm not making important mistakes and to make it as readable as I could. Would it be valuable if it was even less readable and more likely to contain mistakes?

Comment by edoarad on Babbling on Singleton Governance · 2019-07-05T18:58:34.799Z · score: 1 (1 votes) · EA · GW

Hypersonic missiles are An example of a possible game changing weapon.

Comment by edoarad on Upcoming interviews on the 80,000 Hours Podcast · 2019-07-05T07:55:18.076Z · score: 1 (1 votes) · EA · GW

Laura Deming: I'm very interested in the tradeoff and relation between investing more in basic science (with the goal of major theoretical breakthroughs) and developing technology and products (which can have immediate impact but perhaps not as important in the long term).

For example,

1. What is the role of the Academia in progress on aging, as opposed to private research groups? In general, how much resources do they have and what are they incentivized to?

2. What are the current bottlenecks in longevity research and tech?

3. How do you view the market and general public view of longevity in the near future?

Comment by edoarad on Target Malaria begins a first experiment on the release of sterile mosquitoes in Africa · 2019-07-05T05:01:18.451Z · score: 2 (2 votes) · EA · GW

Thanks Vox's Future Perfect newsletter for curating this :)

Comment by edoarad on [deleted post] 2019-06-24T07:55:59.254Z

Seems almost identical to a recent post.

Comment by edoarad on Babbling on Singleton Governance · 2019-06-24T07:53:39.178Z · score: 5 (2 votes) · EA · GW

Actually, it is from Robin Hanson, encountered recently when reading this interesting post from someone who recreated GPT2 and considered publishing the result :)

Comment by edoarad on Which World Gets Saved · 2018-11-09T19:43:11.327Z · score: 8 (5 votes) · EA · GW

Very interesting and well written, thank you for this important idea!

Playing around with toy examples, I noticed another version of this consideration. The success of some actions might imply that there is a higher likelihood of a dangerous world (for which there might be better alternative actions).

For example, consider allocating resources for handling bio-terrorism. One important question is how likely is it that there will be individuals that want to cause massive harm and are able to. If this resource allocation does indeed stop a global extinction, this means that we should update our estimated likelihood for such an individual. In this world, there may better alternatives than focusing on something specific to current biological techniques and instead to put effort into methods which can help against any individual that wants to cause massive harm by any technology.

Comment by edoarad on Why the EA Forum? · 2018-11-08T11:47:02.418Z · score: 12 (7 votes) · EA · GW

One thing which causes me and probably many others to avoid writing more on the forum, is the feeling of writing posts which "spams" or lowers the standards of the forum. This is not mentioned here as a reason not to post. I guess that the voting system and the option of writing on the personal blog solves most of this issue, and that we prefer to encourage more people to write more instead of focusing on quality for now?

Comment by edoarad on On Becoming World-Class · 2018-11-02T15:10:29.021Z · score: 6 (4 votes) · EA · GW

My main takeaway from this post is that people who are on a clear path to becoming world-class (or already are) can be great advocates for EA or for effective giving among their (fringe/prestigious) peer group, and that the EA community should be welcoming and encouraging them to be active members of the community. I do think that the concerns raised and the framing of the conclusion are apt.

Engaging with the prompt, I am considering an academic career. This may not lead to me being world class (say, win a Nobel Prize), but this is a likely path to becoming high status which has similar merits. I guess that in the technical or research oriented world, it is easier to have both high status and to do EA-aligned work, so I do not think of this as a major consideration.

I find it interesting that the three arguments given also apply somewhat to the value of high social skills:

1. Better social skills lead to stronger connections. People with exceptional social skills can also be more likely to connect with high status individuals.

2. Being more charismatic helps one be more persuasive (in a different manner than written above).

3. Having more people with high social skills in the community can help newcomers feel more at ease and engaged with EA.