Saving the world in 80 days: Prologue

post by Logan Riggs (elriggs) · 2018-05-09T21:16:03.875Z · LW · GW · 16 comments

Contents

  What is this?
  Who are you?
  Why?
  So what are you going to be doing exactly?
  Aren't you going to flounder/fail/burn-out?
None
16 comments

What is this?

This is an intro to an intensive productivity sprint for becoming useful in the field of AI safety. I'll be posting every week or 4 on the good/bad/ugly of it all. Definitely inspired by Nate Soares' productivity sprint [LW · GW].

Who are you?

Recent graduate in Computer Engineering who has someone to protect [LW · GW]

Why?

Ultimately to further the field of AI safety. In the process of doing this and writing these posts, I'll:

1. Create a more realistic model of what I can do/ of my limits

2. Receive valuable feedback from other rationalists (that's you!)

3. Radically improve my knowledge of the field & various maths

So what are you going to be doing exactly?

16 hours in a day:

Aren't you going to flounder/fail/burn-out?

I've done 3 other intensive sessions like this in the past. One was a flop and two were quite successful, so I feel prepared for this one. I'm definitely shooting for sustainability here, so I'm trying to figure out my limits without burning out. This could include getting more work done per hour, getting more efficient rest, or varying the amount of slack time given.

My current Murphyjitsu [LW · GW] results:

1. I could waste a lot of time on pica [LW · GW]. To prevent that, I'm going to notice when I want to leave my current situation and pursue that and just watch that thought/sensation without fueling it. I've developed this skill while meditating, and have had some success already, though I feel like there's more that's missing. Something like I'm quickly making deals with myself such as "You only have to deal with this for a few more minutes, then you're free".

2. I could get uber mentally tired. If so, I'll move on to living maintenance or meditation which I consider mentally relaxing.

3. Trips & friends taking a lot of time. I'm not going to go out of my way to spend time with people, but important trips (weddings, family) I'm going to for sure, and I'm okay with less productivity; I'm just gonna' enjoy the heck out of it.

[Note: I've actually been trying to do this intensive session this past week, but apparently people want to spend time with you after you graduate college or something. I even got food poisoning, lol, but I still managed to get 6 hours into my current book! So I'm officially starting tomorrow after I fully recover]

If you have any questions, comments, derogatory remarks, please comment it! I'll check them once in the mornings (probably)

16 comments

Comments sorted by top scores.

comment by Michaël Trazzi (mtrazzi) · 2018-05-10T09:51:49.186Z · LW(p) · GW(p)

Awesome!

The energy from this post is real.

You, sir, just bootstrapped my motivation so here are a few techniques/experiences:

  • Two years ago I wanted to study hard for a really though exam. What I did at the time: I published a status on Facebook saying that I would study 14 hours a day for the next 14 days with my bestfriend, and that if I didn't do it I would pay him 100 euros + give him my favorite sweatshirt. I also published a post on stickk to make the payment official. At that time, I was also waking up everyday before 4:30am for another challenge (similar setting). What I learned: With enough accountability, from (friends, colleagues, family), I can study hard, wake up very early, and do impressive things (when I think about it now, I really had a shit ton of energy at that time). However, mental/physical stretches alone don't necessarily imply that you will achieve something you find meaningful. Yes, I was really productive during this 14 days sprint. But I failed at my exams (my body was exhausted and I got sick two days before) and I didn't even learn anything useful (I studied to pass the exam, and didn't study deeply the material). Practical Advice: 16 hours a day is a long period of time. You might want to shorten this time to maybe 10, or 12 max. I know, when people told me this three years ago, I also thought they were not motivated enough. They were right. Please, try to spend at least 2-4 hours a day doing some really relaxing stuff, go outside, meet people, do something were you're truly uninterested in the results, just wanting to chill. This relaxing time is the most important time of the day, and no, 'I'll move on to living maintenance or meditation which I consider mentally relaxing.' doesn't seem to me like enough relaxation. Meditation is a mental stretch, and living maintenance/food/hygiene(sleeping, shower, food, etc.) does make you more relaxed, but is not really an activity where you can immerse yourself, meet people, talk about your emotions and let off steam. In particular, try not to eat/do living maintenance alone.
  • I used to write down some great plan (like you just did) every month or so, and giving up after one week every time. I remember saying to my friend that I would work more, code at least 4 hours a day, etc. None of this worked for me (in the long term). What I learned: whenever I make a plan (e.g. work 4 hours a day), a) I kind of force myself to do something I don't really want to do. Another way of putting this: I like the result, but I don't like the process (see this book). So, in the long term, I always end up quitting. Furthermore, b) I am always overly optimistic, forgetting life's hazards (planning fallacy). 4 hours a day (in my example) is a lot, and if you're busy all day with something else, you might not even have the time to do it (yesterday's scenario ("So I'm officially starting tomorrow") might happen again). Practical advice: About three weeks ago, I started being absurdly productive, by assigning a color to every 15m of my life on a Google Sheet. For the last five years, I tried to boost my productivity, and this is the technique that worked the best with me so far. Why this works: I don't force myself to do anything. I know what is important to me (writing and seeing people), and, just by looking at the colors, I know if I have been doing random unimportant stuff, or things that I value. In other words, I have enough flexibility to do whatever I like, but it also gives me enough motivation to write for hours, because I emotionally connect with my time. So, as a practical advice, I would say: try to insert more flexibility. This is especially true, given that you said "I'm definitely shooting for sustainability here, so I'm trying to figure out my limits without burning out"

Now, here are some questions about your program:

  • What exactly are you trying to achieve? Ok, you'll be doing some AI Safety reading, and some tensorflow. But like, what is your deeper motivation? Can you state your life-statement in a sentence? You commented on the AISF program in June, do you want to prepare for that? Can you give some more info about "someone to protect"?Feedback: think your post definitely needs a longer introduction (with a real paragraph, not bullet points) about what you're doing and why you're doing it.
  • "Create a more realistic model of what I can do/ of my limits" : what you can do in AI Safety? The impact you can have in the field? Your limits as a Computer Scientist/AI Researcher? Or, are you trying to grasp the limits of your productivity? Your biological limits?
  • "Radically improve my knowledge of the field & various maths" : could you be more specific? What exactly do you fell you're lacking? What would you want to study more? Is there some particular sub-domain that interest you most?

This is quite a long comment, because your post moved me. I also want to deepen my understanding of AI Safety, and also love to challenge my productivity. I will happily comment on your posts. I have been doing a series of daily Less Wrong posts for 12 days in a row, sometimes about AI Safety, sometimes about general AI/Philosophy/Decision Theory, and the energy in this post motivated me to keep writing and to keep pushing my limits. See you on the other side!

Replies from: elriggs, TurnTrout
comment by Logan Riggs (elriggs) · 2018-05-10T21:24:10.485Z · LW(p) · GW(p)

Wow! Glad good things are already coming out of this!

Thanks for sharing your experiences and the warning with it (this is the type of post I'd like to promote!), though I predict I'll do well in this program due to what TurnTrout said in the other comment: I enjoy a lot of what I'm doing! * actually considers each item *... yep! This is honestly what I'd rather be doing than a lot of things, so I feel like Nate Soares in that regard (in his post I linked).

Regarding my why/motivation/someone to protect, I'm going to leave that for a separate post. I wanted this one to be a short & to the point intro. My why post will be much more poetic and wouldn't fit here, and to separate it more cleanly, I'm referring to a terminal goal here.

Though I would love to clarify my instrumental goals to achieve that terminal goal! Those are those 3 bullet points "better self-model, feedback, & self-improvement".

Better self-model: I would like to ~maximize my usefulness which would require working hard for several years (So closest to "productivity/ biological limits"). Getting the most bang for my buck those years involves finding a sustainable sprint/jog, so I'm making predictions and testing those predictions to get a more accurate self-model.

Self-improvement: I feel lacking in math and technical knowledge of open-problems in AI safety (as well as how progress has been made so far).

comment by TurnTrout · 2018-05-10T14:55:57.692Z · LW(p) · GW(p)

I don't think this schedule is overly ambitious in how much time is being dedicated to reading / how "little" to free time. I mean, if we're swapping anecdotes, I probably don't spend more than half an hour on explicit free time per day. Instead, I find ways to enjoy what I'm doing / relax throughout the day, whether it be through music, enjoying some sunshine while I walk and read, or something else entirely.

comment by Mitchell_Porter · 2018-05-09T22:34:56.057Z · LW(p) · GW(p)

An exercise for those who are a little more advanced: *actually* save the world within the next 80 days. In the context of AI safety at the singularity level, that would mean completely figuring out the theory required to have a friendly singularity, and then making it happen for real, by the end of July.

Replies from: ChristianKl, gjm
comment by ChristianKl · 2018-05-10T09:48:53.282Z · LW(p) · GW(p)

I downvoted the comment. We want to have a culture where people who chose to do good ambitious projected to get social approval instead of being told: "What you want to do isn't very advanced, advanced people should do X".

Replies from: Mitchell_Porter
comment by Mitchell_Porter · 2018-05-10T10:29:45.863Z · LW(p) · GW(p)

The fact that I was basically serious, and in no way attempting to discourage @elriggs, and yet the comment is (after 12 hours) at -17, suggests that LW now has a problem with people who *do* want to do advanced things.

Replies from: gjm, ChristianKl
comment by gjm · 2018-05-10T11:25:53.499Z · LW(p) · GW(p)

I don't think it should be at -17, but I don't think what its low score indicates is that LW has a problem with people who want to do advanced things.

Your suggestion, taken as a serious one, is obviously absurdly overambitious. Its premise is that the experts in the relevant fields have nearly enough knowledge to (1) create a superhuman AI and (2) arrange for it to behave in ways that are good rather than bad for us. But so far as I can tell, those same experts are pretty much universally agreed that those are super-hard problems. (There are some people who are arguably experts on #1 and think #2 might be easy. There are some people who are arguably experts on #2 and think #1 might happen much sooner than we'd guess. But I don't think I know of anyone who's an expert on #1 and thinks #1 is feasible within, say, a year, or of anyone who's an expert on #2 and thinks #2 is feasible on that timescale.)

So you are simultaneously arguing that the state of the art in these things is really far advanced and that the people whose work makes it so are hopelessly incompetent to evaluate how close we are.

For sure, it could turn out that you're right. But it seems staggeringly unlikely, and in any case anyone actually in a position to solve those problems within 80 days is surely already working on them.

Also: If I try to think about the possible worlds most similar to the one I think we're actually in in which at least one of those problems does get solved within 80 days, it seems to me that a substantial fraction are ones in which just one of them does, and the most likely way for that to happen is some sort of rapidly recursively self-improving AI ("FOOM"), and if that happens without the other problem getting solved there's a substantial danger that we're all screwed. In that possible world, advising people to rush to solve those problems seems like rather a bad idea.

(I don't think FOOM+doom is a terribly likely outcome. But I think it's quite likely conditional on any part of your proposal turning out to be feasible.)

Replies from: Viliam
comment by Viliam · 2018-05-11T23:56:14.604Z · LW(p) · GW(p)
I don't think it should be at -17

Just a reminder that karma works slightly differently on LW 2.0, so karma -17 today means less than karma -17 would have meant on LW 1.0.

comment by ChristianKl · 2018-05-11T07:42:09.059Z · LW(p) · GW(p)

The problem isn't conscious intent but the social effect of a statement. Being bad at social skills and thus lacking awareness of making status moves is no good justification for them being proper.

If your position is to seriously propose that trial, there's no good reason to do so in this thread but you could have written your own post for it.

comment by gjm · 2018-05-09T22:51:55.881Z · LW(p) · GW(p)

What's the point of the comment above? I mean, presumably you don't actually think that's a sensible goal for anyone to have, because no one could think that, so I guess the purpose is mockery or something of the kind -- but if so, presumably there's some point beyond mere mean-spiritedness, some actual implicit argument that you're making, and I'm not seeing it.

(Perhaps I'm completely misunderstanding and you actually do think it would make sense to take "completely solve the problem of creating superhuman AI, along with the problem of having it not destroy everything we care about, in 80 days" as a goal. In that case, please accept my apologies and feel free to explain why you think such an extraordinary thing.)

Replies from: Mitchell_Porter, jameson-quinn, elriggs
comment by Mitchell_Porter · 2018-05-09T23:26:13.298Z · LW(p) · GW(p)

I do mean it rather seriously. There are theoretical frameworks already in play, directed at solving the two major subproblems that you identify, i.e. creating raw superintelligence, and (let's say) identifying a friendly value system. I actually find it conceivable that the required breakthroughs are not far away, in the same way that e.g. imminent solutions to math's Millennium Problems are conceivable - someone just needs to have the right insights.

Replies from: TurnTrout
comment by TurnTrout · 2018-05-10T00:03:12.044Z · LW(p) · GW(p)

Disclaimer: it's hard to see too many levels above your own. This may still be underestimating the difficulty of friendliness.

I'd like to note that I'm skeptical of this for a few reasons:

1) this doesn't even come close to making sense if you aren't already as knowledgeable as one of the MIRI staff, in which case an 80-day resolve cycle to this effect might actually produce something. I think it is very unlikely that the entire problem can be resolved without an insane amount of intuition (both mathematical and normal). Even if a promising solution was found in that time frame, however,

2) there's insufficient time for verification. Proving your intuitively-appealing idea will be safe is different than making strong arguments it will.

someone just needs to have the right insights.

Saying "there may be a short inferential distance to the full solution" (as if step-count is the main quantifier of difficulty / time required) misses the difficulty that some of these steps may entail. Yes, the challenge may be different than that posed to a rationalist in 1300 who realizes the horror of death and wants to live. Even if he had complete knowledge of what steps to take, it would be incredibly difficult (and probably impossible) for him to single-handedly build the machinery required to advance science beyond today's frontier so as to ensure he continued living. That's for the case in which he spends every waking moment taking an optimal action (with respect to a scientific gradient even more advanced than today's). Yes, in that situation, there are too many insights and actions required for him to survive.

In that sense, yes - the problem is, perhaps, literally able to be solved by one person's efforts, but I still don't think it's a reasonable challenge. Maybe if you're already on or near Eliezer's level, this extended resolve cycle would be useful for generating new ideas.

comment by Jameson Quinn (jameson-quinn) · 2018-05-09T23:09:48.851Z · LW(p) · GW(p)

It's probably just a lame joke that didn't land. I wouldn't downvote it below "a little bit negative" which is where it currently stands.

Replies from: gjm, elriggs
comment by gjm · 2018-05-09T23:29:08.707Z · LW(p) · GW(p)

For what it's worth, I didn't downvote it. I wasn't sure enough of what it's trying to achieve to be able to tell whether it succeeded or failed, or whether I approve or disapprove :-).

comment by Logan Riggs (elriggs) · 2018-05-09T23:23:36.397Z · LW(p) · GW(p)

Agreed on the "not downvoting any more than it is right now (-2)". Though I would still like to dissuade any comments not directly related to the content of the post!

comment by Logan Riggs (elriggs) · 2018-05-09T23:17:50.469Z · LW(p) · GW(p)

I initially interpreted Mitchell's as mocking as well, but on a second...third read I interpreted it as:

A reference to a common text book theme "An exercise for the reader" combined with the title of this post. Meant as a funny-joke-but-would-also-be-really-cool-if-someone-actually-did-it. This is just speculation though!! (Is this 100% correct Mitchell?)

I greatly appreciate you standing up for me though!!

If my speculation is correct, then I think the reason both you and I originally interpreted it as mockery would be the "those who are a little more advanced" part (meant as hyperbole) and the "*actually*" part.