Ulisse Mini's Shortform

post by Ulisse Mini (ulisse-mini) · 2022-05-05T14:21:51.879Z · LW · GW · 28 comments

Contents

28 comments

28 comments

Comments sorted by top scores.

comment by Ulisse Mini (ulisse-mini) · 2022-05-05T14:21:52.721Z · LW(p) · GW(p)

Some realizations about memory and learning I've been thinking about recently EDIT: here are some great posts on memory which are a deconfused version of this shortform (and written by EY's wife!)

  • Anki (and SRS in general) is a tool for efficiently writing directed graph edges to the brain. thinking about encoding knowledge as a directed graph can help with making good Anki cards.
  • Memory techniques are somewhat-analogous to data structures as well, e.g. the link method corresponds to a doubly linked list
  • "Memory techniques" should be called "Memory principles" (or even laws).
  • The "Code is Data" concept makes me realize memorization is more widely applicable, you could e.g. memorize the algorithm for integration in calculus. Many "creative" processes like integration can be reduced to an algorithm.
  • Truely part of you [LW · GW] is not orthogonal to memory techniques principles, it uses the fact that a densely connected graph is less likely to be disconnected from randomly deleting edges, similar to how the link and story methods. Just because you aren't making silly images doesn't mean you aren't using the principles.
  • (untested idea for math): Journal about your thought processes after solving each problem, then generalize to form a problem solving algorithm / checklist and memorize the algorithm
Replies from: MackGopherSena
comment by MackGopherSena · 2022-05-05T16:02:05.781Z · LW(p) · GW(p)

[edited]

Replies from: eigen, ulisse-mini
comment by eigen · 2022-05-05T22:10:50.775Z · LW(p) · GW(p)

Good comments, thanks for sharing both. 

Journal about your thought processes after solving each problem

create a way to 'get to' that memory somehow;

'd love to hear more about practical insights on how to get better at recalling + problem-solving.

Replies from: ulisse-mini
comment by Ulisse Mini (ulisse-mini) · 2022-05-05T23:12:47.485Z · LW(p) · GW(p)

I'll write some posts when I get stuff working, I feel a Sense That More Is Possible [? · GW] in this area, but I don't want to write stuff till I can at least say it works well for me.

comment by Ulisse Mini (ulisse-mini) · 2022-05-05T21:23:00.539Z · LW(p) · GW(p)

= finding shortest paths on a weighted directed graph, where the shortest path cost must be below some threshold :)

comment by Ulisse Mini (ulisse-mini) · 2022-11-06T00:10:02.105Z · LW(p) · GW(p)

In a recent post, [LW · GW] John mentioned how Corrigability being a subset of Human Values means we should consider using Corrigability as an alignment target. This is a useful perspective, but I want to register that  doesn't always imply that doing  is "easier" than , this is similar to the problems with problem factorization for alignment [? · GW] but even stronger! Even if we only want to solve  and not ,  can still be harder!

For a few examples of this:

  • Acquiring half a Strawberry by itself is harder than a full strawberry (You have to get a full strawberry, then cut it in half) (This holds for X=MacBook, Person, Penny too)
  • Let  be a lemma used in a proof of   (meaning  in some sense). It may be that  can be immediately proved via a known more general theorem In this case  is harder to directly prove then .
  • When writing an essay, writing section 3 alone can be harder than writing the whole essay, because it interacts with the other parts, you learn from writing the previous parts, etc. (Sidenote: There's a trivial sense[1] in which writing section 3 can be no harder than writing the whole essay, but in practice we don't care as the whole point to considering a decomposition is to do better.)

In general, depending on how "natural" the subproblem in the factorization is, subproblems can be harder than solving the original problem. I believe this may (30%) be the case with corrigibility; mainly because (1) Corrigability is anti-natural in some ways, and (2) Humans are pretty good at human values while being not-that-corrigible.

  1. ^

    Just write the whole thing then throw everything else away!

Replies from: D0TheMath
comment by Garrett Baker (D0TheMath) · 2022-11-06T00:27:08.883Z · LW(p) · GW(p)

I believe this may (30%) be the case with corrigibility

Surprising agreement with my credence! First skim I thought "Uli isn't thinking correctly about how humans may have an explicit value for corrigible things, so if humans have 10B values, and we have an adequate value theory, solving corrigibility only requires searching for 1 value in the brain, while solving value-alignment requires searching for 10B values", and I decided I thought this class of arguments brings something roughly corresponding to 'corrigibility is easier' to 70%. But then I looked at your credence, and turned out we agreed.

Replies from: ulisse-mini
comment by Ulisse Mini (ulisse-mini) · 2022-11-06T00:44:02.688Z · LW(p) · GW(p)

Mmm, I think it matters a lot which of the 10B[1] values is harder to instill, I think most of the difficulty is in corrigibility. Strong corrigibility seems like it basically solves alignment. If this is the case then corrigibility is a great thing to aim for, since it's the real "hard part" as opposed to random human values. I'm ranting now though... :L


  1. I think it's way less than 10B, probably <1000 though I haven't thought about this much and don't know what you're counting as one "value" (If you mean value shard maybe closer to 10B, if you mean human interpretable value I think <1000) ↩︎

comment by Ulisse Mini (ulisse-mini) · 2022-11-30T19:54:51.368Z · LW(p) · GW(p)

There are a series of math books that give a wide overview of a lot of math. In the spirit of comprehensive information gathering [AF · GW], I'm going to try to spend my "fun math time" reading these.

I theorize this is a good way to build mathematical maturity, at least the "parse advanced math" part. I remember when I became mathematically mature enough to read Math Wikipedia, I want to go further in this direction till I can read math-y papers like Wikipedia.

Replies from: Jay Bailey, jack-o-brien
comment by Jay Bailey · 2022-12-01T15:29:14.298Z · LW(p) · GW(p)

This seems like an interesting idea. I have this vague sense that if I want to go into alignment I should know a lot of maths, but when I ask myself why, the only answers I can come up with are:

  • Because people I respect (Eliezer, Nate, John) seem to think so (BAD REASON)
  • Because I might run into a problem and need more maths to solve it (Not great reason since I could learn the maths I need then)
  • Because I might run into a problem and not have the mathematical concepts needed to even recognise it as solvable or to reduce it to a Reason 2 level problem (Good reason)

I wonder if reading a book or two like that would provide a good amount of benefit towards Reason 3 without requiring years of study.

Replies from: ulisse-mini, jack-o-brien
comment by Ulisse Mini (ulisse-mini) · 2022-12-01T18:28:13.557Z · LW(p) · GW(p)

#3 is good. another good reason is so you have enough mathematical maturity to understand fancy theoretical results.

I'm probably overestimating the importance of #4, really I just like having the ability to pick up a random undergrad/early-grad math book and understand what's going on, and I'd like to extend that further up the tree :)

comment by Jack O'Brien (jack-o-brien) · 2022-12-02T03:11:35.950Z · LW(p) · GW(p)

3 is my main reason for wanting to learn more pure math, but I use 1 and 2 to help motivate me

comment by Jack O'Brien (jack-o-brien) · 2022-12-01T09:02:42.551Z · LW(p) · GW(p)

which of these books are you most excited about and why? I also want to do more fun math reading

Replies from: ulisse-mini
comment by Ulisse Mini (ulisse-mini) · 2022-12-01T18:14:43.614Z · LW(p) · GW(p)

(Note; I haven't finished any of them)

Quantum computing since Democritus is great, I understand Godel's results now! And a bunch of complexity stuff I'm still wrapping my head around.

The Road to Reality is great, I can pretend to know complex analysis after reading chapters 5,7,8 and most people can't tell the difference! Here's a solution to a problem in chapter 7 I wrote up.

I've only skimmed parts of the Princeton guides, and different articles are written by different authors—but Tao's explanation of compactness (also in the book) is fantastic, I don't remember specific other things I read.

Started reading "All the math you missed" but stopped before I got to the new parts, did review linear algebra usefully though. Will definitely read more at some point.

I read some of The Napkin's guide to Group Theory, but not much else. Got a great joke from it:

comment by Ulisse Mini (ulisse-mini) · 2022-11-28T04:07:24.109Z · LW(p) · GW(p)

Here's a collection of links to recent "mind-reading" related results using deep learning. Comment ones I missed!

PS: It seems like a good idea for alignment people (like Steve [AF · GW]) to have the capacity to run novel human-brain experiments like these. If we don't currently have this capacity, well... Free dignity points to be gained :)

Replies from: belkarx
comment by belkarx · 2022-11-28T05:52:41.361Z · LW(p) · GW(p)Replies from: ulisse-mini
comment by Ulisse Mini (ulisse-mini) · 2022-11-28T05:55:24.830Z · LW(p) · GW(p)

You're probably in lesswrong docs mode, either switch to markdown or press ctrl+k to insert link around selected text

Replies from: belkarx
comment by belkarx · 2022-11-28T05:58:19.190Z · LW(p) · GW(p)

Do you know how to get out of docs mode? (nvm - got it - thanks!)

comment by Ulisse Mini (ulisse-mini) · 2022-09-23T21:50:11.511Z · LW(p) · GW(p)

I'm going to argue meditation/introspection skill is a key part of an alignment researcher's repituaire. I'll start with a somewhat fake Taxonomy of approaches to understanding intelligence/agency/value formation

  • Study artificial intelligences
    • From outside (run experiments, predict results, theorize)
    • From inside (interpretability)
  • Study biological intelligences
    • From outside (psychology experiments, theorizing about human value & intelligence formation, study hypnosis[1])
    • From inside (neuroscience, meditation, introspection)

I believe introspection/meditation is a neglected way to study intelligence among alignment researchers

  • You can run experiments on your own mind at any time. Lots of experimental bits free for the taking
  • I expect interviewing high level meditators to miss most of the valueble illegible intuitions (both from lack of direct experience, and lacking the technical knowledge to integrate that experience with)
  • It has known positive side effects like improved focus, reduced stress etc (yes it is possible to wirehead, I believe it to be worth the risk if you're careful though.)

What are you waiting for, get started![2]


  1. I threw this out offhand, unsure if it's a good idea, but maybe figuring out hypnosis will teach us something about the mind? (Also hypnosis could be in the "outside" or "inside" category.) ↩︎

  2. Or read the mind illuminated, mastering the core teachings of the Buddha, joy on demand. Better yet find a teacher. ↩︎

Replies from: niplav
comment by niplav · 2022-10-03T23:02:16.822Z · LW(p) · GW(p)

I disagree with your stated reasons for meditating (I have meditated a bunch and believe it to be valuable, but not worth the time if valence is not one of your top priorities).

Feedback is less clear (not so many experimental bits), experiments are repeated very often (and often at least I don't have the focus to think about better methods), I expect positive side-effects to be very small due to the Algernon argument.

I think I would attribute having become better at the 12th virtue of rationality (the void) due to meditation practice, but this is quite hard to tell. Maybe also better at not fooling myself in social situations, slightly less consequentialist/analytical in macro-thinking (but no noticeable decrease in micro-thinking such as programming), and more adept at understanding æsthetics.

Replies from: ulisse-mini
comment by Ulisse Mini (ulisse-mini) · 2022-10-04T20:53:48.872Z · LW(p) · GW(p)

Thanks for the comment. I agree valence should be the top priority & that cognitive gains are unlikely.

The main thing I was pointing at is "surely there's useful bits about intelligence to be gathered from inside said intelligence, and people don't seem to be putting much effort here", but on reflection the last part seems wrong. Some neuroscientists are enlightened and haven't figured everything out yet.

Your experience is interesting, I also want to get better at the void :)

Replies from: niplav
comment by niplav · 2022-10-04T21:23:44.574Z · LW(p) · GW(p)

It could still be true that on the margin experiential information about intelligence is useful.

I think I buy this.

comment by Ulisse Mini (ulisse-mini) · 2023-07-11T20:23:15.926Z · LW(p) · GW(p)

Quick thoughts on creating a anti-human chess engine [LW(p) · GW(p)].

  1. Use maiachess to get a probability distribution over opponent moves based on their ELO. for extra credit fine-tune on that specific player's past games.
  2. Compute expectiminimax search over maia predictions. Bottom out with stockfish value when going deeper becomes impractical. (For MVP bottom out with stockfish after a couple ply, no need to be fancy.) Also note: We want to maximize (P(win)) not centipawn advantage.
  3. For extra credit, tune hyperparameters via self-play against maia (simulated human). Use lichess players as a validation set.
  4. ???
  5. Profit.
comment by Ulisse Mini (ulisse-mini) · 2022-06-02T21:34:52.123Z · LW(p) · GW(p)

People Power

To get a sense that more is possible [? · GW] consider

  1. The AI box experiment, and its replication [LW · GW]
  2. Mentalists like Derren Brown (which is related to 1 [LW(p) · GW(p)])
  3. How the FBI gets hostages back with zero leverage (they aren't allowed to pay ransoms)

(This is an excerpt from a post I'm writing which I may or may not publish. the link aggregation here might be useful in of itself)

comment by Ulisse Mini (ulisse-mini) · 2023-03-21T22:37:29.940Z · LW(p) · GW(p)

Petition to rename "noticing confusion" to "acting on confusion" or "acting to resolve confusion". I find myself quite good at the former but bad at the latter—and I expect other rationalists are the same.

For example: I remember having the insight thought leading to lsusr's post [LW · GW] on how self-reference breaks the orthogonality thesis, but never pursued the line of questioning since it would require sitting down and questioning my beliefs with paper for a few minutes, which is inconvenient and would interrupt my coding.

comment by Ulisse Mini (ulisse-mini) · 2023-02-11T00:52:12.766Z · LW(p) · GW(p)

Exercise: What mistake is the following sentiment making?

If there's only a one in a million chance someone can save the world, then there'd better be well more than a million people trying.

Answer:

The whole challenge of "having a one in a million chance of saving the world" is the wrong framing, the challenge is having a positive impact in the first case (for example: by not destroying the world or making things worse, e.g. from s-risks). You could think of this as a setting the zero point [LW · GW] thing going on, though I like to think of it in terms of Bayes and Pascel's wagers:

In terms of Bayes: You're fixating on the expected value contributed from  and ignoring the rest of the  hypothesis space. In most cases, there are corresponding low probability events which "cancel out" the EV contributed from 's direct reasoning.

(I will also note that, empirically, it could be argued Eliezer was massively net-negative from a capabilities advancements perspective; having causal links to founding of deepmind & openai. I bring this up to point out how nontrivial having a positive impact at all is, in a domain like ours)

comment by Ulisse Mini (ulisse-mini) · 2022-12-29T17:50:18.749Z · LW(p) · GW(p)

Are most uncertainties we care about logical rather than informational? All empirical ML experiments are pure computations a Bayesian superintelligence could do in its head. How much of our uncertainty comes from computational limits in practice, versus actual information bottlenecks?

comment by Ulisse Mini (ulisse-mini) · 2022-07-28T19:30:33.962Z · LW(p) · GW(p)

KL-divergence and map territory distinction

Crosspost from my blog


The cross-entropy is defined as the expected surprise when drawing from , which we're modeling as . Our map is while is the territory.

Now it should be intuitively clear that because an imperfect model will (on average) surprise us more than the perfect model .

To measure unnecessary surprise from approximating by we define

This is KL-divergence! The average additional surprise from our map approximating the territory.

Now it's time for an exercise, in the following figure is the Gaussian that minimizes or , can you tell which is which?

Left is minimizing while the right is minimizing .

Reason as follows:

  • If is the territory then the left is a better map (of ) than the right .
  • If is the map, then the territory on the right leads to us being less surprised than the territory on the left, because on the on left will be very surprised at data in the middle, despite it being likely according to the territory .

On the left we fit the map to the territory, on the right we fit the territory to the map.