Posts
Comments
I agree with whay you say about how to maximize what you get out of an interview. I also agree about that discussion vs. debate distinction you make, and I wasn't specifically trying to go there when I used the word "debate", I was just sloppy with words.
I guess you agree that it is friction to create a social norm that you should do a read up of the other person material before engaging in public. I expect less discussions would happen. There is not a clear threshold at how much you should be prepared.
I guess we disagree about how much value do we lose due to eliminating discussions that could have happaned, vs. how much value we gain by eliminating some lower quality discussions.
Another angle I have in mind that sidesteps this direct compromise, is that maybe what we value out of such discussions is not just doing an optimal play in terms of information transmitted between the parties. A public discussion has many different viewers. In the case at hand, I expect many people get more out of the discussion if they can see Wolfram think through the thing for the first time in real time, rather than having two informed people start discussing finer points in medias res.
I see your proposed condition for meaningful debate as bureaucracy that adds friction rather than value.
I somewhat disagree with Tenobrus' commentary about Wolfram.
I watched the full podcast, and my impression was that Wolfram uses a "scientific hat", of which he is well aware of, which comes with a certain ritual and method for looking at new things and learning them. Wolfram is doing the ritual of understanding what Yudkowsky says, which involves picking at the details of everything.
Wolfram often recognizes that maybe he feels like agreeing with something, but "scientifically" he has a duty to pick it apart. I think this has to be understood as a learning process rather than as a state of belief.
So, should the restrictions on gambling be based on feedback loop length? Should sport betting be broadly legal when about the far enough future?
current inference scaling methods tend to be tied to CoT and the like, which are quite transparent
Aschenbrenner in Situational Awareness predicts illegible chains of thought are going to prevail because they are more efficient. I know of one developer claiming to do this (https://platonicresearch.com/) but I guess there must be many.
Related, I have a vague understanding on how product safety certification works in EU, and there are multiple private companies doing the certification in every state.
Half-informed take on "the SNPs explain a small part of the genetic variance": maybe the regression methods are bad?
Not sure if I missed something because I read quickly, but: all these are purely correlational studies, without causal inference, right?
OpenAI is recklessly scaling AI. Besides accelerating "progress" toward mass extinction, it causes increasing harms. Many communities are now speaking up. In my circles only, I count seven new books critiquing AI corps. It’s what happens when you scrape everyone's personal data to train inscrutable models (computed by polluting data centers) used to cheaply automate out professionals and spread disinformation and deepfakes.
Could you justify that it causes increasing harms? My intuition is that OpenAI is currently net-positive without taking into account future risks. It's just an intuition, however, I have not spent time thinking about it and writing down numbers.
(I agree it's net-negative overall.)
Ok, that. China seems less interventionist, and to use more soft power. The US is more willing to go to war. But is that because the US is more powerful than China, or because Chinese culture is intrinsically more peaceful? If China made the killer robots first, would they say "MUA-HA-HA actually we always wanted to shoot people for no good reason like in yankee movies! Go and kill!"
Since politics is a default-no on lesswrong, I'll try to muddle the waters by making a distracting unserious figurative narration.
Americans maybe have more of a culture of "if I die in a shooting conflict, I die honorably, guns for everyone". Instead China is more about harmony&homogenity, "The CCP is proud to announce that in 2025 the Harmonious Agreement Quinquennal Plan in concluded successfully; all disagreements are no more, and everyone is officially friends". When the Chinese send Uighurs to the adult equivalent of school, Americans freak out: "What? Mandated school? Without the option of shooting back?"
My doubt is mostly contingent on not having first-hand experience of China, while I have of the US. I really don't trust narratives from outside. In particular I don't trust narratives from Americans right now! My own impression of the US changed substantially by going there in person, and I even am from an allied country with broad US cultural influence.
[Alert: political content]
About the US vs. China argument: have any proponent made a case that the Americans are the good guys here?
My vague perspective as someone not in China neither in the US, is that the US is overall more violent and reckless than China. My personal cultural preference is for US, but when I think about the future of humanity, I try to set aside what I like for myself.
So far the US is screaming "US or China!" while creating the problem in the first place all along. It could be true that if China developed AGI it would be worse, but that should be argued.
I bet there is some more serious non-selfish analysis of why China developing AGI is worse than US developing AGI, I just have never encountered it, would be glad if someone surfaced it to me.
I agree it's not a flaw in the grand scheme of things. It's a flaw for using it for consensus for reasoning.
I start with a very low prior of AGI doom (for the purpose of this discussion, assume I defer to consensus).
You link to a prediction market (Manifold's "Will AI wipe out humanity before the year 2100", curretly at 13%).
Problems I see with using it for this question, in random order:
- It ends in 2100 so the incentive is effectively about what people will believe a few years from now, not about the question. It is a Keynesian beauty contest. (Better than nothing.)
- Even with the stated question, you win only if it resolves NO, so it is strategically correct to bet NO.
- It is dynamically inconsistent, if you think that humans have power over the outcome and that such markets influence what humans do about it. Illustrative story: "The market says P(doom)=1%, ok I can relax and not work on AI safety" => everyone says that => the market says P(doom)=99% because no AI safety work => "AAAAH SOMEONE DO SOMETHING" => marker P(doom)=1% => ...
This type of issue is a huge effective blocker for people with my level of skills. I find myself excited to write actual code that does the things, but the thought of having to set everything up to get to that point fills with dread – I just know that the AI is going to get something stupid wrong, and everything’s going to be screwed up, and it’s going to be hours trying to figure it out and so on, and maybe I’ll just work on something else. Sigh. At some point I need to power through.
Reminds me of this 2009 kalzumeus quote:
I want to quote a real customer of mine, who captures the B2C mindset about installing software very eloquently: “Before I download yet another program to my poor old computer, could you let me know if I can…” Painful experience has taught this woman that downloading software to her computer is a risky activity. Your website, in addition to making this process totally painless, needs to establish to her up-front the benefits of using your software and the safety of doing so. (Communicating safety could be an entire article in itself.)
Ah, sorry for being so cursory.
A common trope about mathematicians vs. other math users is that mathematicians are paranoid persnickety truth-seekers, they want everything to be exactly correct down to every detail. Thus engineers and physicists often perceive mathematicians as a sort of fact-checker caste.
As you say, in some sense mathematicians deal with made-up stuff and engineers with real stuff. But from the engineer's point of view, they deal with mathematicians when writing math, not when screwing bolts, and so perceive mathematicians as "the annoying people who want everything to be perfectly correct".
Example: I write "E[E[X|Y]] = E[X]" in a paper, and the mathematician pops up complaining "What's the measure space? Is it sigma-finite? You have to declare if your random variables are square-integrable. Are X and Y measureable in the same space?" and my reply would be "come on we know it's true I don't care about writing it properly".
So to me and many people in STEM your analogy has the opposite vibe, which defeats the purpose of an analogy.
The analogy with mathematicians is very stretched.
- Include, in the cue to each note, a hint as to its content, besides just the ordinal pointer. A one-letter abbreviation, standardised thruout the work, may work well, e.g.:
- "c" for citation supporting the marked claim
- "d" for a definition of the marked term
- "f" for further, niche information extending the marked section
- "t" for a pedantic detail or technicality modifying the marked clause
- Commit to only use notes for one purpose — say, only definitions, or only citations. State this commitment to the reader.
These don't look like good solutions to me. Just a first impression.
I don't make eye contact while speaking but fix people while in silence. Were there people like me? Did they managed to reverse this? The way I feel inside is more like I can't think both about the face of someone and what I am saying at once, too many things to keep track of.
Is champerty legal in California?
I take this as a fun occasion to lose some of my karma in a silly way to remind myself lesswrong karma is not important.
A very interesting problem is measuring something like general intelligence. I’m not going to delve deeply into this topic but simply want to draw attention to an idea that is often implied, though rarely expressed, in the framing of such a problem: the assumption that an "intelligence level," whatever it may be, corresponds to some inherent properties of a person and can be measured through their manifestations. Moreover, we often talk about measurements with a precision of a few percentage points, which suggests that, in theory, the measurement should be based on very stable indicators.
What’s fascinating is that this assumption receives very little scrutiny, while in cases where we talk about "mechanical" parameters of the human body (such as physical performance), we know that such parameters, aside from a person’s potential, heavily depends on numerous external factors and what that person has been doing over the past couple of weeks.
Do you know about the g-factor?
Still 20$/usd
Does it imply that normally everyone would receive as many spam calls, but the more expensive companies are spending a lot of their budget to actively fight against the spammers?
Yeah, they said this is what happens.
"things that mysteriously don't work in USA despite working more or less okay in most developed countries"
Let my try:
- expensive telephone (despite the infamous bell breakup)
- super-expensive health
- super-expensive college
- no universally free or dirty cheap wire transfers between any bank
- difficult to find a girlfriend
- no gun control (disputed)
- crap train transport
- threatening weirdos on the street
- people are bad at driving
- no walkable town center
- junk fees
- tipping loses more value in attrition than the incentive alignment it creates (disputed)
- overly complicated way to pay at restaurants
- people still use checks
- limited power out of home power outlets due to 110V and small plugs/wires
- low quality home appliances
- long queue at TSA (it could actually take only 1 min)
I don't really know about this specific proposal to deter spam calls, but speaking in general: I'm from another large first world country, and when staying in the US a striking difference was receiving on average 4 spam calls per day. My american friends told me it was because my phone company was low-cost, but it was O(10) more expensive (per unit data) than what I had back home, with about O(1) spam calls per year.
So I expect that it is totally possible to solve this problem without doing something too fancy, even if I don't know how it's solved where I am from.
Most people do not have the analytical clarity to be able to give an explanation of love isomorphic to their implementation of love; to that extent, they are "confused about love".
This though does not imply that their usage of the word "love" is amiss, the same way people are able to get through simple reasoning without learning logic, or walking without learning Physics.
So I'll assume that people are wielding "love" meaningfully, and try to infer what the word means.
It seems to indicate prolonged positive emotional involvement with an external entity. Other emotions and reactions surrounding this may be dragged in. As everything emotions, the boundaries are not clearly defined, but categories were made for man, not the reverse.
This degree of vagueness may be unnerving if you are more detail-oriented than most people. It is familiarly unnerving to me, as I am pretty detail-oriented. (To readers that are not detail-oriented, a concrete example of being so: searching for a big red thing in store shelves, you notice the big red thing, I scan sequentially the items until I hit it.)
Then the question becomes: Is the category "love" too broad?
First line of argument: if you can put 4 items in two sets of 2 items, you can also put them in a set of 4 items, even if this offends the aestethic sense of @SpectrumDT. This is not a good line of argument, as there is probably a more grounded meaning to the question; it's here just to put it out of the way.
Second line of argument: do some inner mental representations actually match such a broad category? Is this category an unnatural herding of many internal sensations, or is there actually something going on in the mind that naturally brings everything together, leading to humans spontaneously uttering a "love" token in response?
I'll attack this question from the front and from behind.
From the front, humans in general tend naturally to quash things into low-precision numbers in their head, because Von Neumann was right. I know how many favors and tricks you have played on me, and feel if you come out positive. I meet a girl, and as I get to know her, how nice she looks changes based on how intelligent she looks. So it's reasonable that the brain tends to summarize a lot of stuff into a level for "I'm emotionally attached in some positive way to that external entity".
From behind, if humans empirically insist on using this concept, it's probably there for a reason. There must be some naturalness to it.
Third line of argument: would people be better of if, by default, they outclassed their monkey brain in the degree of accuracy they use to think about love? Here I think the answer depends on the intellectual abilities of a person. If you are intelligent enough, at some point it's advantageous to make more complex categories and models. Below some level of intelligence, though, the result of trying to install into someone a more complex love model may not be worth the effort. That person may be better off with a long list of proverbs and heuristics that involve less things at a time, and more broad things, such that pattern matching is easier. Detailed discernment is reached by majority/importance vote over multiple loose patterns.
Conclusion: I overall think love is an adequate abstraction for most people, but not for a minority of detail-oriented and intelligent ones.
(Note: I have the impression the writing style I just used gives a vibe of "I know what I am saying, I'm a PhD in love analysis". I am not.)
I'm reminded of the recent review of How Language Began on ACX: the missionary linguist becomes an atheist because in the local very weird language they have declinations to indicate the source of what you are saying, and saying things about Jesus just doesn't click.
I still don't understand your "infinite limit" idea. If in your post I drop the following paragraph:
A way to think about the proposition is as a kind of limit. When we have little evidence, each bit of evidence has a potentially big impact on our overall probability of a given proposition. But each incremental bit of evidence shifts our beliefs less and less. The proposition can be thought of a shorthand for an infinite collection of evidences where the collection leads to an overall probability of given to . This would perhaps explains why the proposition is so strange: we have well-developed intuitions for how "finite" propositions interact, but the characteristic absorbing property of the distribution is more reminiscent of how an infinite object interacts with finite objects.
the rest is standard hierarchical modeling. So even if your words here are suggestive, I don't understand how to actually connect the idea to calculations/concrete things, even at a vague indicative level. So I guess I'm not actually understanding it.
For example, you could show me a conceptual example where you do something with this which is not standard probabilistic modeling. Or maybe it's all standard but you get to a solution faster. Or anything where applying the idea produces something different, then I would see how it works.
(Note: I don't know if you noticed, but De Finetti applies to proper infinite sequences only, not finite ones, people forget this. It is not relevant to the discussion though)
I happened to have the same doubt as you. A deeper analysis of the sacred texts shows how your interpretation of the Golden Rule is amiss. You say:
“Do unto others as you would have them do unto you.” (Matthew 7:12)
Therefore whatever you desire for men to do to you, you shall also do to them; for this is the law and the prophets.
The verse speaks specifically of men, not generically of others. So if you are straight, it does not compel you to sexual acts on women, while if you are gay, you shall try to hit on all the men to your heart's content. You can see this perfectly fits ordinary morality.
I think it's apposite to go beyond the Golden Rule, but by improvement rather than negation. The first step is Kant's Categorical Imperative: "Act only according to that maxim whereby you can at the same time will that it should become a universal law.". The second step is functional decision theory.
There are too many nonpolar bears in the US to keep up the lie.
I guess the point of the official party line is to avoid kids going and trying to scare polar bears.
I don't think this concept is useful.
What you are showing with the coin is a hierarchical model over multiple coin flips, and doesn't need new probability concepts. Let be the flips. All you need in life is the distribution . You can decide to restrict yourself to distributions of the form . In practice, you start out thinking about as a variable atop all the in a graph, and then think in terms of and separately, because that's more intuitive. This is the standard way of doing things. All you do with is the same, there's no point at which you do something different in practice, even if you ascribed additional properties to in words.
A concept like "the probability of me assigning a certain probability" makes sense but I don't think Jaynes actually did anything like that for real. Here on lesswrong I guess @abramdemski knows about stuff like that.
--PS: I think Jaynes was great in his way of approaching the meaning and intuition of statistics, but the book is bad as a statistics textbook. It's literally the half-complete posthumous publication of a rambling contrarian physicist, and it shows. So I would not trust any specific statistical thing he does. Taking the general vibe and ideas is good, but when you ask about a specific thing "why is nobody doing this?" it's most likely because it's outdated or wrong.
A group of MR links led to a group of links that led to this list of Obvious Travel Advice. It seems like very good Obvious Travel Advice, and I endorse almost all points.
> A place that has staff trying to flag down customers walking past is almost certainly pursuing a get people in the door strategy.
To my great surprise, I found this to be false in Pisa (n_restaurant = 2).
Timothy Bates: The more things change, the more they stay the same: 1943 paper shows that a mechanical prediction of admissions greatly out predicts the decisions from administrators asked to add their subjective judgement :-(excellent talk from Nathan Kuncel !)
Nick Brown: I would bet that if you asked those subjective evaluators, they would say “We know the grades are the best predictor on average, but ‘sometimes’ they don’t tell the whole story”. People want to double-dip: Use the method most of the time, but add their own “special expertise”.
Timothy Bates: Nathan Kuncel put it astutely showing that decision makers beta weights are pretty accurate, but then they ruin their decision at “run time” by adding random intuitions about details in the application :-)
[in the figure: algorithm r(rating, GPA) = 0.45), humans r(rating, GPA) = 0.35
I don't like that this reasoning is based on using the correlation between rating and GPA, because I think GPA is goodharted. It is right to not select the admission process based on correlation with GPA. I think maybe this would be the case even if the humans where adding white noise.
I joke about calling them ‘Openly Evil AI,’ but developing a 99.9% effective watermarking tool and then sitting on it because people identifying your outputs would be bad for business? Yeah, that’s something.
Maybe if you solve for equilibrium you get that after releasing the tool, the tool is defeated reasonably quickly?
[json formatting in gpt]
How did they do it?
> OpenAI: While sampling, after every token, our inference engine will determine which tokens are valid to be produced next based on the previously generated tokens and the rules within the grammar that indicate which tokens are valid next. We then use this list of tokens to mask the next sampling step, which effectively lowers the probability of invalid tokens to 0. Because we have preprocessed the schema, we can use a cached data structure to do this efficiently, with minimal latency overhead.
Throw out all invalid outputs, and all the outputs that remain will be valid. Nice.
This is obvious. Why wasn't it available already? I guess bandwidth is what it is.
In the "Agent Performance vs Humans with Time Limits (95% CI)" figure: the 95% cl bars are fishy, because they look large compared to the regularity of the bars. i.e., the bars smoothly match the relative rankings I already expected from the models: if there were independent fluctuations of that size, the bars would be jagged compared to the expected ranking based on general model quality we already know. Possible explanations:
- They made a mistake and the bars are based on the population sdev rather than the standard error. In this case the actual bars would be smaller.
- The LLM scores are correlated: the randomness is given by the tasks, not by the models, so there's some uncertainty related to the small number of tasks, but the inter-LLM comparisons are pretty accurate. In this case, the bars for comparing LLMs would be smaller, while the present bars are fine for comparing LLM to humans (probably).
Unpolished first thoughts:
- Selection effect: people who go to a blog to read bc they like reading, not doing
- Concrete things are hard reads, math-heavy posts, doesn't feel ok to vote when you don't actually understand
- In general easier things have wider audience
- Making someone change their mind is more valuable to them than saying you did something?
- There are many small targets and few big ideas/frames, votes are distributed proportionally
Talking 1-1 with music is so difficult to me that I don't enjoy a place if there's music. I expect many people on/towards the spectrum could be similar.
I didn't catch this thing was not from an official anthropic doc. I think you should add something to the title or the first paragraph to clarify this, e.g., "Re: Anthropic's suggested SB-1047 amendments (unofficial)"
Learning Universal Predictors explores what happens if you take ideas from Solomonoff Induction and train actual neural network architectures like transformers on data generated from Universal Turing Machines.
I already found this here somehow, I can't recall where
Ok, let's say the clone is mirrored.
For (1), think about this concrete toy example: say the first thing you do is raising your right hand. The mirror clone rises his left hand. Then say you have an instinctive reaction behaving in this way:
- if I see a person raising his right hand, I repeat that movement.
- if I see a person raising his left hand, I do nothing.
So the next thing you do is nothing, while the next thing the mirror clone does is raising his right hand, which differentiates you from the clone.
HOWEVER, I now realize I was wrong: the mirror clone's notion of left-right is mirrored as well.
- As a human, you react to what other humans do around you. Your reaction algorithms may not allow two copies of you to mirror each other.
- The clone is not mirrored, so you are not in front of a mirror you, but a rotated you.
Anna Tong and Katie Paul (Reuters): The document describes a project that uses Strawberry models with the aim of enabling the company’s AI to not just generate answers to queries but to plan ahead enough to navigate the internet autonomously and reliably to perform what OpenAI terms “deep research,” according to the source.
Is this by any chance inspired by Yudkowsky's "clone a strawberry" example?
If you give it proper labels, an LLM can learn that some information (e.g. Wikipedia) is reliable and should be internalized, whereas others (e.g. 4chan) is unreliable and should only be memorized.
Do the labels give the ground truth about which is reliable, or does the LLM learn which is the reliable one?
Before doing any project or entering any field, you need to catch up on existing intellectual discussion on the subject.
My current take on this topic is to follow this scheme:
1) dedicate some time to think about the problem on your own, without searching the literature
2) look in the literature, compare with your own thoughts
3) get feedback from a human in the field
4) repeat
Do you think (1) makes sense, or is your position extreme enough to reject (1) altogether, or spend only a very short time on it, say < 1 hour?
Isn't it the opposite? To decimate = to kill 1 in 10 soldiers, figuratively to remove a certain fraction of elements from a set.
Simply run experiments and accept every result as true if the probability of it occurring by random chance falls below some threshold we agree on. This will lead us terribly astray every once in a while if we are not careful, but it also enables us to run experiments whose conclusions both of us can trust.[2]
To minimize the chance of statistical noise or incorrect inference polluting our conclusions, we create experiments with randomly chosen intervention and control groups, so we are sure the intervention is causally connected to the outcome.
As long as we follow these procedures exactly, we can both trust the conclusion. Others can even join in on the fun too.
Together we arrive at a set of ‘randomista’ interventions we both recognize as valuable. Even if we each have differing priors leading us to opposing preferred interventions, pooling our money together on the randomista interventions beats donating to causes which cancel each other out.
The world is some the better.
Problems I see here:
- In an RCT you can get a causal effect measure rather than just a correlation, but you can't prove whether it generalizes. Example: If you pick some people and flip a coin for each to decide whether to give them a blerg or a sblarg, and then find out that blergs has a positive effect on self-confidence, the weak spot in your inference is the way you have picked the initial set of people within whom you did the randomization. The debate moves from "you can't prove it's a causal effect!" to "your set of people where younger than normal", "the set of people is from a first-world country", "people changed too much from when you did the experiment to right now", etc.
- Considering only significance and not power (in other words, only checking the probability something would happen "by chance", rather than the probability it would happen under specific meaningful alternative hypotheses) is limiting and can be misleading enough to be a problem in practice. Read Andrew Gelman for this stuff.
- By Wald's complete class theorems, whatever you do with kosher frequentist statistics that you think is good, is equivalent to some specific prior used by a Bayesian guy doing Bayesian things, so if you at any point think "but frequentist" you should also have a caveat like "considering that humans won't actually use statistics kosher, I can see a pretty precise story for how doing nonsense in this specific way will produce the correct behavior in the end", which incidentally is reasoning I would not trust people in general to pull off.
As Ian notes, in theory one could consider ASML or TSMC as more plausible monopolies, but neither is exploiting its position, and also neither is American so we can’t go after them. If anything I find the continued failure of both to raise prices to be a confusing aspect of the world.
Maybe with something that grows exponentially it's convenient to keep the prices lower to grow the pie faster?
Indeed, I have spoken how I feel proposals such as those in the Gladstone Report go too far, and would indeed carry exactly these risks. I draw a sharp contrast between that and something like SB 1047. I dive into the details to try and punch them up.
I'm not a native English speaker, could someone tell me what "punch smth up" means here? I couldn't understand searching on wiktionary nor asking claude.
I didn’t play along and ended up saying things like: “Thanks, but I have friends I can talk to if I need it”.
Do you think the default thing to say in these cases should assume you have any friends period, or any "friends" you are willing to speak to?
Naia: It’s OK, everyone. Mr. Not Consistently Candid says the whole thing was an oopsie, and that he’ll fix things one-by-one for people if they contact him privately. Definitely nothing to worry about, then. Carry on.
I'm wondering, what would happen if you contact Altman privately right now? Would you be added to a list of bad kids? What is the typical level of shadiness of American VCs?
Look, I've measured hands with women, and not personally done it out of a relationship but I've seen women resting their heads and purring, and I don't think it was flirting. Either this or I totally don't understand social relationships (possible!) Do you think this is cultural? I'm Italian, are you US American?
Ok, then I agreed. I was interpreting the advice in a different way, but your interpretation looks more reasonable.
Thinking about it, I suspect I was not getting what "authenticity and openness" means. Like, it's not "being yourself and letting go", and more "being honest", I guess? Could you give me >= 2 examples of a person being "authentic and open"?
Are you libertarian about this specifically? Do you think it's better if people also have the choice of dating apps? Or would you ban them if given the choice?