Posts

Aligning Mathematical Notions of Infinity with Human Intuition 2023-06-12T19:19:07.931Z

Comments

Comment by London L. (london-l) on Aligning Mathematical Notions of Infinity with Human Intuition · 2023-06-20T22:16:42.316Z · LW · GW

If the color of the number is considered to be an intrinsic property of the number, then under the Bruce Framework, yes, |C|<|B| and |C|=|A| and |B|=|A|.

Comment by London L. (london-l) on Aligning Mathematical Notions of Infinity with Human Intuition · 2023-06-20T22:11:54.234Z · LW · GW

So then because the winner alternates at an even rate between the two sets, you can intuitionally guess that they are equal?

Comment by London L. (london-l) on Aligning Mathematical Notions of Infinity with Human Intuition · 2023-06-13T06:13:41.765Z · LW · GW

I like this a lot! I'm curious, though, in your head, what are you doing when you're considering an "infinite extent of "? My guess is that you're actually doing something like the "markers" idea (though I could be wrong), where you're inherently matching the extent of  on A to the extent of  on B for smaller-than-infinity numbers, and then generalizing those results.

For example, when thinking through your example of alternating pairs, I'm checking to see when =3, that's basically containing the 2 and everything lower, so I mark 3 and 2 as being the same, and then I do the density calculation. Matching 3 to 2 and then 7 to 6, I see that each set always has 2 elements in each section, so I conclude that they have an equal number of elements.

Does this "matching" idea make sense? Do you think it's what you do? If not, what are your mental images or concepts like when trying to understand what happens at the "infinite extent"? (I imagine you're not immediately drawing conclusions from imagining the infinite case, and are instead building up something like a sequence limit or pattern identification among lower values, but I could be wrong.)

Comment by London L. (london-l) on Aligning Mathematical Notions of Infinity with Human Intuition · 2023-06-13T06:06:14.329Z · LW · GW

Yep, absolutely! It was actually through explaining Hilbert's Hotel that Bruce helped me come up with the Bruce Framework.

I do think it is odd though that the mathematical notion of cardinality doesn't solve the Thanos Problem, and I'm worried that AI systems that understand math practically well but not theoretically well will consider the loss of half an infinite set to be no loss at all, similar to how if you understand Hilbert's you'll believe that adding twice the number of hotels is never an issue.

Comment by London L. (london-l) on Aligning Mathematical Notions of Infinity with Human Intuition · 2023-06-12T19:32:55.121Z · LW · GW

I'm posting this here because I find that I don't get the feedback or discussion that I want in order to improve my ideas on Medium. So I hope that people leave comments here so we can discuss this further.

Personally, I've come across two other models of how humans intuitively compare infinities.

One of them is that humans use a notion of "density". For example, positive multiples of three (3, 6, 9, 12, etc.) seem like a smaller set than all positive numbers (1, 2, 3, etc.). You could use the Bruce Framework here, but I think that what we're actually doing something closer to evaluating the density of the sets. We notice that 3 and 6 and 9 are in both sets (similar to the Bruce Framework), but then we look to see how many numbers are between those "markers". In the positive numbers set, there are 3 numbers between each "marker" (3, 4, 5 and then 6, 7, 8), whereas in the set of positive multiples of three there is only 1 number between each "marker" (3 and then we immediately go to 6). Thus, the cardinality of the positive numbers must be 3 times bigger than the cardinality of positive multiples of three.

If you expand this sort of thinking further, you get to a more "meta-model" of how humans intuitively compare sets, which is that we seem to build simple and easy functions to map items in one set to items in the other set. Sometimes the simple function is "are these inherently equal", as in the Bruce Framework. Other times it's "obvious" function like converting a negative number to a positive number. Once we have this mapping of "markers", we then use density to compare the sizes of the two sets.

I'm not 100% sure if density is the only intuitive metric we use, but from the toy examples in my head it is. What are your thoughts? Are there any infinite sets (numbers or objects or anything) where your intuitive calculation doesn't involve pairing up markers between the sets and then evaluating the density between those markers?

Comment by London L. (london-l) on Models Don't "Get Reward" · 2023-02-22T00:37:49.609Z · LW · GW

To (rather gruesomely) link this back to the dog analogy, RL is more like asking 100 dogs to sit, breeding the dogs which do sit and killing those which don't.  Overtime, you will have a dog that can sit on command. No dog ever gets given a biscuit.

The phrasing I find most clear is this: Reinforcement learning should be viewed through the lens of selection, not the lens of incentivisation.


I was talking through this with an AGI Safety group today, and while I think the selection lens is helpful and helps illustrate your point, I don't think the analogy quoted above is accurate in the way it should be.

The analogy you give is very similar to genetic algorithms, where models that get high reward are blended together with the other models that also get high reward and then mutated randomly. The only process that pushes performance to be better is that blending and mutating, which doesn't require any knowledge of how to improve the models' performance.

In other words, carrying out the breeding process requires no knowledge about what will make the dog more likely to sit. You just breed the ones that do sit. In essence, not even the breeder "gets the connection between the dogs' genetics and sitting".

However, at the start of your article, you describe gradient descent, which requires knowledge about how to get the models to perform better at the task. You need the gradient (the relationship between model parameters and reward at the level of tiny shifts to the model parameters) in order to perform gradient descent.

In gradient descent, just like the genetic algorithms, the model doesn't get access to the gradient or information about the reward. But you still need to know the relationship between behavior and reward in order to do the update. In essence, the model trainer "gets the connection between the models' parameters and performance".

So I think a revised version of the dog analogy that takes this connection information into account might be something more like:
1. Asking a single dog to sit, and monitoring its brain activity during and afterwards.
2. Search for dogs that have brain structures that you think would lead to them sitting, based on what you observed in the first dog's brain.
3. Ask one of those dogs to sit and monitor its brain activity.
4. Based on what you find, search for a better brain structure for sitting, etc.
At the end, you'll have likely found a dog that sits when you ask it to.

A more violent version would be:
1. Ask a single dog to sit and monitor brain activity
2. Kill the dog and modify its brain in a way you think would make it more likely to sit. Also remove all memories of its past life. (This is equivalent to trying to find a dog with a brain that's more conducive to sitting.)
3. Reinsert the new brain into the dog, ask it to sit again, and monitor its brain activity.
4. Repeat the process until the dog sits consistently.

To reiterate, with your breeding analogy, the person doing the breeding doesn't need to know anything about how the dogs' brains relate to sitting. They just breed them and hope for the best, just like in genetic algorithms. However, with this brain modification analogy, you do need to know how the dog's brain relates to sitting. You modify the brain in a way that you think will make it better, just like in gradient descent.

I'm not 100% sure why I think that this is an important distinction, but I do, so I figured I'd share it, in hopes of making the analogy less wrong.

Comment by London L. (london-l) on Models Don't "Get Reward" · 2023-02-21T23:54:57.673Z · LW · GW

When you say "the dog metaphor" do you mean the original one with the biscuit, or the later one with the killing and breeding?

Comment by London L. (london-l) on How it feels to have your mind hacked by an AI · 2023-01-18T02:02:51.977Z · LW · GW

It is! You (and others who agree with this) might be interested in this competition (https://futureoflife.org/project/worldbuilding-competition/) which aims to create more positive stories of AI, which may help shift pop culture in a positive direction.

Comment by London L. (london-l) on How it feels to have your mind hacked by an AI · 2023-01-18T01:58:55.104Z · LW · GW

I had a friend in a class today where you need to know the programming language C in order to do the class. But now with ChatGPT available, I told them it probably wasn't that big of an issue, as you could probably have ChatGPT teach you C as you go through the class. I probably would have told them they should drop the class just one semester ago (before ChatGPT).

My personal analogy has been that these chat bots are like a structural speed up for humans in a similar way that Google Docs and Drive were for working on documents and files with people - it's a free service that everyone just has access to now to talk through ideas or learn things. It's ethical to use, and if you don't use it, you'll probably not be as capable as those who do.

Comment by London L. (london-l) on AGI Ruin: A List of Lethalities · 2022-09-17T23:24:11.149Z · LW · GW

Small typo in point (-2): "Less than fifty percent change" --> "Less than 50 percent chance"

Comment by London L. (london-l) on Lies Told To Children · 2022-04-18T11:41:30.264Z · LW · GW

I say this because I think it relates to / could be within your same world:

One of my favorite current story ideas is to write a story where there's a girl who lives in a world where each society is running an experiment about the best way to live life. In school, she finds out that while most societies are aware of the experiment that's being run, her society is one of the few where the experiment is kept secret from the people in it.

As she grows up, she realizes how poor her society is, and how little technology they have. It improves very quickly over the years, but when she was younger they didn't even have washing machines or dishwashers.

So, she starts digging to try to figure out why it is that all their technology was so bad, and why her society started off so poor, and begins advocating for the idea that whatever experiment is being run, it should be shut down, because their quality of life is so much lesser than other societies.

Eventually, the movement she starts begins gaining momentum, and the entire society begins to revolt against being in the experiment. When she's finally able to bring her case to the main governing decision body (the council), she finds out what the experiment was:

Despite their wealth, many of the people in the other societies are depressed and feel like they have very little to live for. Is it better to live wealthy and securely without worries, or to struggle but have a purpose of overcoming those challenges?

The later chapters cover the (now older) woman fighting with her emotions. She is proud that she was able to fight for her society to raise its standards, and she feels like she's lived a good life. However, not having the wealth at first was really tough, and she's worried the council will make more societies like hers and subject more people to those tribulations she experienced when she was younger. But, because she and others in her society ended up greatly enjoying their lives despite (and because of, as the experiment showed) the hardships, the experiment was considered a success. If she likes her life overall, should she really oppose the creation of more societies like hers?

Comment by London L. (london-l) on On infinite ethics · 2022-04-14T04:49:53.165Z · LW · GW

I have not read through this in its entirety, but it strikes me that an article I wrote about how the mathematical definition of infinity doesn't match human intuitions might be useful for people to read who are also interested in this material. I'm also fairly new here, so if cross posting this isn't okay, please let me know.

https://london-lowmanstone.medium.com/comparing-infinities-e4a3d66c2b07

Comment by London L. (london-l) on Embedded Interactive Predictions on LessWrong · 2022-03-30T04:17:35.327Z · LW · GW

Is it possible to hide the values of other predictors? I'm worried that seeing the values that others predict might influence future predictions in a way that's not helpful for feedback or accurate group predictions. (Especially since names are shown and humans tend to respect others' opinions.)

Comment by London L. (london-l) on Naturalism · 2022-03-15T05:14:29.062Z · LW · GW

Yeah, that sounds right! I think this video supports that idea as well:

Comment by London L. (london-l) on Naturalism · 2022-02-28T23:15:21.493Z · LW · GW

For example, when asked to think about something I would like more deeper, masterful knowledge about, I replied "artificial neural networks".

The closest thing I could think to potentially experiencing them and interacting is either 1. Through interactive demos or 2. Through a suit like this. I'm unsure if that is what is meant by interaction, though it does seem closer.

Comment by London L. (london-l) on Direct Observation · 2022-02-28T22:58:18.689Z · LW · GW

This is my first comment on this site, so if I'm missing particular norms, please let me know.
 

I understand why this was done, but it is amusing to me that in order to describe what "contact with the territory" looks like, you must use map-like terms such as "arm", "finger", and "hand".

I was going to recommend that you could qualify all of these terms, but I realized that this would likely not be needed for most readers; hopefully most people will understand, after having reading the previous essays, that by "hand", you're taking advantage of our preconceptions and built-in maps in order to describe your thought.

(I also realize that this has been done elsewhere out of necessity in the essays as well - it just particularly stood out to me in this one section.)