Hiring engineers and researchers to help align GPT-3

post by Paul_Christiano · 2020-10-01T18:52:21.088Z · score: 106 (51 votes) · EA · GW · 19 comments

Contents

  What the team does
  What you’d do
None
19 comments

My team at OpenAI, which works on aligning GPT-3, is hiring ML engineers and researchers. Apply here for the ML engineer role and here for the ML researcher role.

GPT-3 is similar enough to "prosaic" AGI that we can work on key alignment problems without relying on conjecture or speculative analogies. And because GPT-3 is already being deployed in the OpenAI API, its misalignment matters to OpenAI’s bottom line — it would be much better if we had an API that was trying to help the user instead of trying to predict the next word of text from the internet.

I think this puts our team in a great place to have an impact:

We have the real-world problems to motivate alignment research, the financial support to hire more people, and a research vision to execute on. We are bottlenecked by excellent researchers and engineers who are excited to work on alignment.

What the team does

In the past Reflection focused on fine-tuning GPT-3 using a reward function learned from human feedback. Our most recent results are here, and had the unusual virtue of simultaneously being exciting enough to ML researchers to be accepted at NeurIPS while being described by Eliezer [LW(p) · GW(p)] as “directly, straight-up relevant to real alignment problems.”

We’re currently working on three things:

If successful, ideas will eventually move up this list, from the conceptual stage to ML prototypes to real deployments. We’re viewing this as practice for integrating alignment into transformative AI deployed by OpenAI or another organization.

What you’d do

Most people on the team do a subset of these core tasks:

If you are excited about this work, apply here for the ML engineer role and here for the ML researcher role.

19 comments

Comments sorted by top scores.

comment by Halstead · 2020-10-07T08:37:29.614Z · score: 25 (13 votes) · EA(p) · GW(p)

Very interesting role, but my understanding was that job posts were not meant to be posted on the forum

comment by Paul_Christiano · 2020-10-09T16:37:46.088Z · score: 22 (10 votes) · EA(p) · GW(p)

My process was to check the "About the forum" link on the left hand side, see that there was a section on "What we discourage" that made no mention of hiring, then search for a few job ads posted on the forum and check that no disapproval was expressed in the comments of those posts.

comment by Aaron Gertler (aarongertler) · 2020-10-07T18:38:57.750Z · score: 6 (4 votes) · EA(p) · GW(p)

That's not my understanding. As the lead moderator, here's what I've told people who ask about job posts:

If we start to have a lot of them such that they're getting in the way of more discussion-ready content, I'd want to keep them off the frontpage. Right now, we only get them very occasionally, and I'm generally happy to have them be more visible...

...especially if it's a post like this one which naturally leads to a bunch of discussion of an org's actual work. (If the job were something like "we need a copyeditor to work on grant reports," it's less likely that good discussion follows, and I'd again consider sorting the content differently.)

If I said something at some point that gave you a different impression of our policy here, my apologies!

comment by Halstead · 2020-10-08T12:58:42.761Z · score: 12 (7 votes) · EA(p) · GW(p)

Before the revamp of the forum, I was asked to take down job ads, but maybe things have changed since then. I personally don't think it would be good for the forum to become a jobs board, since the community already has several places to post jobs.

comment by Aaron Gertler (aarongertler) · 2020-10-09T02:26:48.494Z · score: -11 (7 votes) · EA(p) · GW(p)

I think our policy has been pretty consistent since the revamp (I wasn't around before then), but it's plausible that the previous policy led people not to post many jobs. 

I also don't want the Forum to become a job board, but I think an occasional post along the lines of this one seems fine; I'm neither an engineer nor a researcher, but I found it interesting to learn what OpenAI was up to.

comment by Aaron Gertler (aarongertler) · 2020-10-09T19:11:49.065Z · score: 2 (1 votes) · EA(p) · GW(p)

I see that people don't seem to like the policy very much (or maybe think it wasn't handled consistently before). If anyone who downvoted sees this, would you mind explaining what you didn't like? Do you think we should simply prohibit all job posts, or make sure they never show up on the front page?

comment by Larks · 2020-10-10T03:56:58.515Z · score: 16 (8 votes) · EA(p) · GW(p)

I didn't downvote, but I could imagine someone thinking Halstead had been 'tricked' - forced into compliance with a rule that was then revoked without notifying him. If he had been notified he might have wanted to post his own job adverts in the last few years.

Personally I share your intuitions that the occasional interesting job offer is good, but I don't know how this public goods problem could be solved. No job ads might be the best solution, for all that I enjoyed this one.

comment by Ben Pace · 2020-10-07T18:31:58.348Z · score: 2 (3 votes) · EA(p) · GW(p)

Yeah. Well, not that they cannot be posted, but that they will not be frontpaged by the mods, and instead kept in the personal blog / community section, which has less visibility.

Added: As it currently says on the About page:

Community posts [? · GW]

Posts that focus on the EA community itself are given a "community" tag. By default, these posts will be hidden from the list of posts on the Forum's front page. You can change how these posts are displayed by using...

comment by Halstead · 2020-10-08T13:00:12.020Z · score: 9 (7 votes) · EA(p) · GW(p)

Ok, the post is still labelled as 'front page' in that case, which seems like it should be changed

comment by Aaron Gertler (aarongertler) · 2020-10-09T19:03:28.976Z · score: 4 (2 votes) · EA(p) · GW(p)

To clarify, Halstead, "Community" is now a tag, not a category on the level of "Frontpage". Posts tagged "Community" will still either be "Frontpage" or "Personal Blog".

comment by Aaron Gertler (aarongertler) · 2020-10-09T19:14:11.010Z · score: 2 (1 votes) · EA(p) · GW(p)

This comment is a bit out of date (though I think it was made before I made this edit). The current language is:

Posts that focus on the EA community itself are given a "community" tag. By default, these posts will have a weighting of "-25" on the Forum's front page (see below), appearing only if they have a lot of upvotes. 

We don't hide all "community" posts by default, but they will generally be less prominent on the front page unless a user changes the weighting themselves.

comment by shaybenmoshe · 2020-10-03T20:15:06.258Z · score: 16 (9 votes) · EA(p) · GW(p)

Thank you for posting this, Paul. I have questions about two different aspects.

In the beginning of your post you suggest that this is "the real thing" and that these systems "could pose an existential risk if scaled up".
I personally, and I believe other members of the community, would like to learn more about your reasoning.
In particular, do you think that GPT-3 specifically could pose an existential risk (for example if it falls into the wrong hands, or scaled up sufficiently)? If so, why, and what is a plausible mechanism by which it poses an x-risk?

On a different matter, what does aligning GPT-3 (or similar systems) mean for you concretely? What would the optimal result of your team's work look like?
(This question assumes that GPT-3 is indeed a "prosaic" AI system, and that we will not gain a fundamental understanding of intelligence by this work.)

Thanks again!

comment by Paul_Christiano · 2020-10-05T19:43:28.252Z · score: 19 (9 votes) · EA(p) · GW(p)

I think that a scaled up version of GPT-3 can be directly applied to problems like "Here's a situation. Here's the desired result. What action will achieve that result?" (E.g. you can already use it to get answers like "What copy will get the user to subscribe to our newsletter?" and we can improve performance by fine-tuning on data about actual customer behavior or by combining GPT-3 with very simple search algorithms.)

I think that if GPT-3 was more powerful then many people would apply it to problems like that. I'm concerned that such systems will then be much better at steering the future than humans are, and that none of these systems will be actually trying to help people get what they want.

A bunch of people have written about this scenario and whether/how it could be risky. I wish that I had better writing to refer people to. Here's a post I wrote last year [LW · GW] to try to communicate what I'm concerned about.

comment by shaybenmoshe · 2020-10-05T21:32:21.454Z · score: 8 (6 votes) · EA(p) · GW(p)

Thanks for the response.
I believe this answers the first part, why GPT-3 poses an x-risk specifically.

Did you or anyone else ever write what aligning a system like GPT-3 looks like? I have to admit that it's hard for me to even have a definition of being (intent) aligned for a system GPT-3, which is not really an agent on its own. How do you define or measure something like this?

comment by shaybenmoshe · 2020-10-03T20:18:40.689Z · score: 5 (4 votes) · EA(p) · GW(p)

Quick question - are these positions relevant as remote positions (not in the US)?

(I wrote this comment separately, because I think it will be interesting to a different, and probably smaller, group of people than the other one.)

comment by Paul_Christiano · 2020-10-05T19:34:57.666Z · score: 7 (4 votes) · EA(p) · GW(p)

Hires would need to be able to move to the US.

comment by weeatquince · 2020-10-04T16:08:42.738Z · score: 3 (2 votes) · EA(p) · GW(p)

Hi, quick question, not sure this is the best place for it but curious:
 

Does work to "align GTP-3" include work to identify the most egregious uses for GTP-3 and develop countermeasures?

Cheers

comment by Paul_Christiano · 2020-10-05T19:34:29.258Z · score: 21 (9 votes) · EA(p) · GW(p)

No, I'm talking somewhat narrowly about intent alignment, i.e. ensuring that our AI system is "trying" to do what we want. We are a relatively focused technical team, and a minority of the organization's investment in safety and preparedness.

The policy team works on identifying misuses and developing countermeasures, and the applied team thinks about those issues as they arise today.

comment by evelynciara · 2020-10-07T22:20:53.158Z · score: 0 (2 votes) · EA(p) · GW(p)

Hi Paul, I messaged you privately.