Alexander Gietelink Oldenziel's Shortform

post by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2022-11-16T15:59:54.709Z · LW · GW · 355 comments

Contents

355 comments

355 comments

Comments sorted by top scores.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-27T12:52:22.928Z · LW(p) · GW(p)

My mainline prediction scenario for the next decades.

My mainline prediction * :

  • LLMs will not scale to AGI. They will not spawn evil gremlins or mesa-optimizers. BUT Scaling laws will continue to hold and future LLMs will be very impressive and make a sizable impact on the real economy and science over the next decade. EDIT: since there is a lot of confusion about this point. BY LLM I mean the paradigm of pre-trained transformers. This does not include different paradigms that follow pre-trained transformers but are still called large language models.  
  • there is a single innovation left to make AGI-in-the-alex sense work, i.e. coherent, long-term planning agents (LTPA) that are effective and efficient in data sparse domains over long horizons. 
  • that innovation will be found within the next 10-15 years
  • It will be clear to the general public that these are dangerous 
  • governments will act quickly and (relativiely) decisively to  bring these agents under state-control. national security concerns will dominate. 
  • power will reside mostly with governments AI safety institutes and national security agencies. In so far as divisions of tech companies are able to create LTPAs they will be effectively nationalized. 
  • International treaties will be made to constrain AI, outlawing the development of LTPAs by private companies. Great power competition will mean US and China will continue developing LTPAs, possibly largely boxed. Treaties will try to constrain this development with only partial succes (similar to nuclear treaties). 
  • LLMs will continue to exist and be used by the general public
  • Conditional on AI ruin the closest analogy is probably something like the Cortez-Pizarro-Afonso takeovers [LW · GW]. Unaligned AI will rely on human infrastructure and human allies for the earlier parts of takeover - but its inherent advantages in tech, coherence, decision-making and (artificial) plagues will be the deciding factor.
  •  The world may be mildly multi-polar. 
    • This will involve conflict between AIs.
    • AIs very possible may be able to cooperate in ways humans can't. 
  • The arrival of AGI will immediately inaugurate a scientific revolution. Sci-fi sounding progress like advanced robotics, quantum magic, nanotech, life extension, laser weapons, large space engineering, cure of many/most remaining diseases will become possible within two decades of AGI, possibly much faster. 
  • Military power will shift to automated manufacturing of drones &  weaponized artificial plagues. Drones, mostly flying will dominate the battlefield. Mass production of drones and their rapid and effective deployment in swarms will be key to victory.

 

Two points on which I differ with most commentators: (i) I believe AGI is a real (mostly discrete) thing , not a vibe, or a general increase of improved tools. I believe it is inherently agenctic. I don't think spontaneous emergence of agents is impossible but I think it is more plausible agents will be built rather than grown. 

(ii) I believe in general the ea/ai safety community is way overrating the importance of individual tech companies vis a vis broader trends and the power of governments. I strongly agree with Stefan Schubert's take here on the latent hidden power of government: https://stefanschubert.substack.com/p/crises-reveal-centralisation

Consequently, the ea/ai safety community is often myopically focusing on boardroom politics that are relativily inconsequential in the grand scheme of things. 

*where by mainline prediction I mean the scenario that is the mode of what I expect. This is the single likeliest scenario. However, since it contains a large number of details each of which could go differently, the probability on this specific scenario is still low. 

Replies from: steve2152, dmurfet, ryan_greenblatt, thomas-kwa, Seth Herd, D0TheMath, lcmgcd, James Anthony
comment by Steven Byrnes (steve2152) · 2024-05-29T01:50:40.400Z · LW(p) · GW(p)

governments will act quickly and (relativiely) decisively to  bring these agents under state-control. national security concerns will dominate.

I dunno, like 20 years ago if someone had said “By the time somebody creates AI that displays common-sense reasoning, passes practically any written test up including graduate-level,  (etc.), obviously governments will be flipping out and nationalizing AI companies etc.”, to me that would have seemed like a reasonable claim. But here we are, and the idea of the USA govt nationalizing OpenAI seems a million miles outside the Overton window.

Likewise, if someone said “After it becomes clear to everyone that lab leaks can cause pandemics costing trillions of dollars and millions of lives, then obviously governments will be flipping out and banning the study of dangerous viruses—or at least, passing stringent regulations with intrusive monitoring and felony penalties for noncompliance etc,” then that would also have sounded reasonable to me! But again, here we are.

So anyway, my conclusion is that when I ask my intuition / imagination whether governments will flip out in thus-and-such circumstance, my intuition / imagination is really bad at answering that question. I think it tends to underweight the force compelling goverments to continue following longstanding customs / habits / norms? Or maybe it’s just hard to predict and these are two cherrypicked examples, and if I thought a bit harder I’d come up with lots of examples in the opposite direction too (i.e., governments flipping out and violating longstanding customs on a dime)? I dunno. Does anyone have a good model here?

Replies from: ryan_greenblatt, Lblack
comment by ryan_greenblatt · 2024-05-29T02:46:39.623Z · LW(p) · GW(p)

One strong reason to think the AI case might be different is that US national security will be actively using AI to build weapons and thus it will be relatively clear and salient to US national security when things get scary.

Replies from: steve2152, johnvon
comment by Steven Byrnes (steve2152) · 2024-06-03T12:56:35.665Z · LW(p) · GW(p)

For one thing, COVID-19 obviously had impacts on military readiness and operations, but I think that fact had very marginal effects on pandemic prevention.

For another thing, I feel like there’s a normal playbook for new weapons-development technology, which is that the military says “Ooh sign me up”, and (in the case of the USA) the military will start using the tech in-house (e.g. at NRL) and they’ll also send out military contracts to develop the tech and apply it to the military. Those contracts are often won by traditional contractors like Raytheon, but in some cases tech companies might bid as well.

I can’t think of precedents where a tech was in wide use by the private sector but then brought under tight military control in the USA. Can you?

The closest things I can think of is secrecy orders (the US military gets to look at every newly-approved US patent and they can choose to declare them to be military secrets) and ITAR (the US military can declare that some area of tech development, e.g. certain types of high-quality IR detectors that are useful for night vision and targeting, can’t be freely exported, nor can their schematics etc. be shared with non-US citizens).

Like, I presume there are lots of non-US-citizens who work for OpenAI. If the US military were to turn OpenAI’s ongoing projects into classified programs (for example), those non-US employees wouldn’t qualify for security clearances. So that would basically destroy OpenAI rather than control it (and of course the non-USA staff would bring their expertise elsewhere). Similarly, if the military was regularly putting secrecy orders on OpenAI’s patents, then OpenAI would obviously respond by applying for fewer patents, and instead keeping things as trade secrets which have no normal avenue for military review.

By the way, fun fact: if some technology or knowledge X is classified, but X is also known outside a classified setting, the military deals with that in a very strange way: people with classified access to X aren’t allowed to talk about X publicly, even while everyone else in the world does! This comes up every time there’s a leak, for example (e.g. Snowden). I mention this fact to suggest an intuitive picture where US military secrecy stuff involves a bunch of very strict procedures that everyone very strictly follows even when they kinda make no sense.

(I have some past experience with ITAR, classified programs, and patent secrecy orders, but I’m not an expert with wide-ranging historical knowledge or anything like that.)

comment by johnvon · 2024-05-30T13:01:21.733Z · LW(p) · GW(p)

'when things get scary' when then? 

comment by Lucius Bushnaq (Lblack) · 2024-06-05T14:09:55.754Z · LW(p) · GW(p)

But here we are, and the idea of the USA govt nationalizing OpenAI seems a million miles outside the Overton window.
 

Registering that it does not seem that far out the Overton window to me anymore. My own advance prediction of how much governments would be flipping out around this capability level has certainly been proven a big underestimate. 


 

comment by Daniel Murfet (dmurfet) · 2024-05-28T00:59:03.488Z · LW(p) · GW(p)

I think this will look a bit outdated in 6-12 months, when there is no longer a clear distinction between LLMs and short term planning agents, and the distinction between the latter and LTPAs looks like a scale difference comparable to GPT2 vs GPT3 rather than a difference in kind.  At what point do you imagine a national government saying "here but no further?".

Replies from: cubefox
comment by cubefox · 2024-05-28T17:59:35.767Z · LW(p) · GW(p)

So you are predicting that within 6-12 months, there will no longer be a clear distinction between LLMs and "short term planning agents". Do you mean that agentic LLM scaffolding like Auto-GPT [LW · GW] will qualify as such?

Replies from: dmurfet
comment by Daniel Murfet (dmurfet) · 2024-05-29T01:28:11.421Z · LW(p) · GW(p)

I think scaffolding is the wrong metaphor. Sequences of actions, observations and rewards are just more tokens to be modeled, and if I were running Google I would be busy instructing all work units to start packaging up such sequences of tokens to feed into the training runs for Gemini models. Many seemingly minor tasks (e.g. app recommendation in the Play store) either have, or could have, components of RL built into the pipeline, and could benefit from incorporating LLMs, either by putting the RL task in-context or through fine-tuning of very fast cheap models.

So when I say I don't see a distinction between LLMs and "short term planning agents" I mean that we already know how to subsume RL tasks into next token prediction, and so there is in some technical sense already no distinction. It's a question of how the underlying capabilities are packaged and deployed, and I think that within 6-12 months there will be many internal deployments of LLMs doing short sequences of tasks within Google. If that works, then it seems very natural to just scale up sequence length as generalisation improves.

Arguably fine-tuning a next-token predictor on action, observation, reward sequences, or doing it in-context, is inferior to using algorithms like PPO. However, the advantage of knowledge transfer from the rest of the next-token predictor's data distribution may more than compensate for this on some short-term tasks.

Replies from: sharmake-farah
comment by Noosphere89 (sharmake-farah) · 2024-09-17T17:40:28.283Z · LW(p) · GW(p)

I think o1 is a partial realization of your thesis, and the only reason it's not more successful is because the compute used for GPT-o1 and GPT-4o were essentially the same:

https://www.lesswrong.com/posts/bhY5aE4MtwpGf3LCo/openai-o1 [LW · GW]

And yeah, the search part was actually quite good, if a bit modest in it's gains.

Replies from: alexander-gietelink-oldenziel, dmurfet
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-09-19T14:12:25.111Z · LW(p) · GW(p)

 As far as I can tell Strawberry is proving me right: it's going beyond pre-training and scales inference - the obvious next step. 

A lot of people said just scaling pre-trained transformers would scale to AGI. I think that's silly and doesn't make sense. But now you don't have to believe me - you can just use OpenAIs latest model. 

The next step is to do efficient long-horizon RL for data-sparse domains.  

Strawberry working suggest that this might not be so hard. Don't be fooled by the modest gains of Strawberry so far. This is a new paradigm that is heading us toward true AGI and superintelligence. 

comment by Daniel Murfet (dmurfet) · 2024-09-18T04:12:38.664Z · LW(p) · GW(p)

Yeah actually Alexander and I talked about that briefly this morning. I agree that the crux is "does this basic kind of thing work" and given that the answer appears to be "yes" we can confidently expect scale (in both pre-training and inference compute) to deliver significant gains.

I'd love to understand better how the RL training for CoT changes the representations learned during pre-training. 

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-09-19T14:22:17.578Z · LW(p) · GW(p)

in my reading, Strawberry is showing that indeed scaling just pretraining transformers will *not* lead to AGI. The new paradigm is inference-scaling - the obvious next step is doing RL on long horizons and sparse data domains. I have been saying this ever since gpt-3 came out. 

For the question of general intelligence imho the scaling is conceptually a red herring: any (general purpose) algorithm will do better when scaled. The key in my mind is the algorithm not the resource, just like I would say a child is generally intelligent while a pocket calculator is not even if the child can't count to 20 yet. It's about the meta-capability to learn not the capability. 

As we spoke earlier - it was predictable that this was going to be the next step. It was likely it was going to work, but there was a hopeful world in which doing the obvious thing turned out to be harder. That hope has been dashed - it suggests longer horizons might be easy too. This means superintelligence within two years is not out of the question. 

Replies from: sharmake-farah
comment by Noosphere89 (sharmake-farah) · 2024-09-20T20:00:13.634Z · LW(p) · GW(p)

We have been shown that this search algorithm works, and we not yet have been shown that the other approaches don't work.

Remember, technological development is disjunctive, and just because you've shown that 1 approach works, doesn't mean that we have been shown that only that approach works.

Of course, people will absolutely try to scale this one up now that they found success, and I think that timelines have definitely been shortened, but remember that AI progress is closer to a disjunctive scenario than conjunctive scenario:

I agree with this quote below, but I wanted to point out the disjunctiveness of AI progress:

As we spoke earlier - it was predictable that this was going to be the next step. It was likely it was going to work, but there was a hopeful world in which doing the obvious thing turned out to be harder. That hope has been dashed - it suggests longer horizons might be easy too. This means superintelligence within two years is not out of the question.

https://gwern.net/forking-path

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-09-23T13:50:21.328Z · LW(p) · GW(p)

strong disagree. i would be highly surprised if there were multiple essentially different algorithms to achieve general intelligence*. 

I also agree with the Daniel Murfet's quote. There is a difference between a disjunction before you see the data and a disjunction after you see the data. I agree AI development is disjunctive before you see the data - but in hindsight all the things that work are really minor variants on a single thing that works. 

*of course "essentially different" is doing a lot of work here. some of the conceptual foundations of intelligence haven't been worked out enough (or Vanessa has and I don't understand it yet) for me to make a formal statement here. 

Replies from: sharmake-farah
comment by Noosphere89 (sharmake-farah) · 2024-09-23T19:47:34.988Z · LW(p) · GW(p)

Re different algorithms, I actually agree with both you and Daniel Murfet in that conditional on non-reversible computers, there is at most 1-3 algorithms to achieve intelligence that can scale arbitrarily large, and I'm closer to 1 than 3 here.

But once reversible computers/superconducting wires are allowed, all bets are off on how many algorithms are allowed, because you can have far, far more computation with far, far less waste heat leaving, and a lot of the design of computers is due to heat requirements.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-09-26T14:11:01.678Z · LW(p) · GW(p)

Reversible computing and superconducting wires seem like hardware innovations. You are saying that this will actually materially change the nature of the algorithm you'd want to run?

I'd bet against. I'd be surprised if this was the case. As far as I can tell everything we have so seen so far points to a common simple core of general intelligence algorithm (basically an open-loop RL algorithm on top of a pre-trained transformers). I'd be surprised if there were materially different ways to do this. One of the main takeaways of the last decade of deep learning process is just how little architecture matters - it's almost all data and compute (plus I claim one extra ingredient, open-loop RL that is efficient on long horizons and sparse data novel domains)

I don't know for certain of course. If I look at theoretical CS though the universality of computation makes me skeptical of radically different algorithms. 

comment by ryan_greenblatt · 2024-05-27T17:42:16.281Z · LW(p) · GW(p)

I'm a bit confused by what you mean by "LLMs will not scale to AGI" in combination with "a single innovation is all that is needed for AGI".

E.g., consider the following scenarios:

  • AGI (in the sense you mean) is achieved by figuring out a somewhat better RL scheme and massively scaling this up on GPT-6.
  • AGI is achieved by doing some sort of architectural hack on top of GPT-6 which makes it able to reason in neuralese for longer and then doing a bunch of training to teach the model to use this well.
  • AGI is achieved via doing some sort of iterative RL/synth data/self-improvement process for GPT-6 in which GPT-6 generates vast amounts of synthetic data for itself using various tools.

IMO, these sound very similar to "LLMs scale to AGI" for many practical purposes:

  • LLM scaling is required for AGI
  • LLM scaling drives the innovation required for AGI
  • From the public's perspective, it maybe just looks like AI is driven by LLMs getting better over time and various tweaks might be continuously introduced.

Maybe it is really key in your view that the single innovation is really discontinuous and maybe the single innovation doesn't really require LLM scaling.

comment by Thomas Kwa (thomas-kwa) · 2024-05-27T18:32:13.534Z · LW(p) · GW(p)

I think a single innovation left to create LTPA is unlikely because it runs contrary to the history of technology and of machine learning. For example, in the 10 years before AlphaGo and before GPT-4, several different innovations were required-- and that's if you count "deep learning" as one item. ChatGPT actually understates the number here because different components of the transformer architecture like attention, residual streams, and transformer++ innovations were all developed separately. 

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-27T19:30:01.226Z · LW(p) · GW(p)

I mostly regard LLMs = [scaling a feedforward network on large numbers of GPUs and data] as a single innovation.

Replies from: thomas-kwa
comment by Thomas Kwa (thomas-kwa) · 2024-05-27T20:31:35.321Z · LW(p) · GW(p)

Then I think you should specify that progress within this single innovation could be continuous over years and include 10+ ML papers in sequence each developing some sub-innovation.

comment by Seth Herd · 2024-05-27T21:51:36.916Z · LW(p) · GW(p)

Agreed on all points except a couple of the less consequential, where I don't disagree.

Strongest agreement: we're underestimating the importance of governments for alignment and use/misuse. We haven't fully updated from the inattentive world hypothesis [LW · GW]. Governments will notice the importance of AGI before it's developed, and will seize control. They don't need to nationalize the corporations, they just need to have a few people embedded at theh company and demand on threat of imprisonment that they're kept involved with all consequential decisions on its use. I doubt they'd even need new laws, because the national security implications are enormous. But if they need new laws, they'll create them as rapidly as necessary. Hopping borders will be difficult, and just put a different government in control.

Strongest disagreement: I think it's likely that zero breakthroughs are needed to add long term planning capabilities to LLM-based systems, and so long term planning agents (I like the terminology) will be present very soon, and  improve as LLMs continue to improve. I have specific reasons for thinking this. I could easily be wrong, but I'm pretty sure that the rational stance is "maybe". This maybe advances the timelines dramatically.

Also strongly agree on AGI as a relatively discontinuous improvement; I worry that this is glossed over in modern "AI safety" discussions, causing people to mistake controlling LLMs for aligning the AGIs we'll create on top of them. AGI alignment requires different conceptual work.

comment by Garrett Baker (D0TheMath) · 2024-05-27T18:13:05.697Z · LW(p) · GW(p)

Do you think the final big advance happens within or with-out labs?

Replies from: alexander-gietelink-oldenziel
comment by lukehmiles (lcmgcd) · 2024-05-29T07:05:53.758Z · LW(p) · GW(p)

So somebody gets an agent which efficiently productively indefinitely works on any specified goal, then they just let the government find out and take it? No countermeasures?

comment by James Anthony · 2024-05-28T17:33:38.531Z · LW(p) · GW(p)

What "coherent, long-term planning agents" means, and what is possible with these agents, is not clear to me. How would they overcome lack of access to knowledge, as was highlighted by F.A. Hayek in "The Use of Knowledge in Society"? What actions would they plan? How would their planning come to replace humans' actions? (Achieving control over some sectors of battlefields would only be controlling destruction, of course, it would not be controlling creation.) 

Some discussion is needed that recognizes and takes into account differences among governance structures. What seems the most relevant to me are these cases: (1) totalitarian governments, (2) somewhat-free governments, (3) transnational corporations, (4) decentralized initiatives. This is a new kind of competition, but the results will be like with major wars: Resilient-enough groups will survive the first wave or new groups will re-form later, and ultimately the competition will be won by the group that outproduces the others. In each successive era, the group that outproduces the others will be the group that leaves people the freest. 

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-11-09T17:56:17.085Z · LW(p) · GW(p)

How to prepare for the coming Taiwan Crisis? Should one short TSMC? Dig a nuclear cellar?

Metaculus gives a 25% of a fullscale invasion of Taiwan within 10 years and a 50% chance of a blockade. It gives a 65% chance that if China invades Taiwan before 2035 the US will respond with military force. 

Metaculus has very strong calibration scores (apparently better than prediction markets). I am inclined to take these numbers as the best guess we currently have of the situation. 

Is there any way to act on this information?

Replies from: weibac, weibac
comment by Milan W (weibac) · 2024-11-10T20:56:23.407Z · LW(p) · GW(p)

Come to think of it, I don't think most compute-based AI timelines models (e.g. EPOCH's) incorporate geopolitical factors such as a possible Taiwan crisis. I'm not even sure whether they should. So keep this in mind while consuming timelines forecasts I guess?

comment by Milan W (weibac) · 2024-11-10T21:21:21.736Z · LW(p) · GW(p)

The CSIS wargamed a 2026 Chinese invasion of Taiwan, and found outcomes ranging from mixed to unfavorable for China (CSIS report). If you trust both them and Metaculus, then you ought to update downwards on your estimate of the PRC's strategic ability. Personally, I think Metaculus overestimates the likelihood of an invasion, and is about right about blockades.

Replies from: ChristianKl, D0TheMath
comment by ChristianKl · 2024-11-11T09:48:21.206Z · LW(p) · GW(p)

Why would you trust CSIS here? A US think tank like that is going to seek to publically say that invading Taiwan is bad for the Chinese.

Replies from: weibac
comment by Milan W (weibac) · 2024-11-11T18:51:47.599Z · LW(p) · GW(p)

Why would they? It's not like the Chinese are going to believe them. And if their target audience is US policymakers, then wouldn't their incentive rather be to play up the impact of marginal US defense investment in the area?

comment by Garrett Baker (D0TheMath) · 2024-11-10T21:46:19.629Z · LW(p) · GW(p)

If you trust both them and Metaculus, then you ought to update downwards on your estimate of the PRC's strategic ability.

I note that the PRC doesn't have a single "strategic ability" in terms of war. They can be better or worse at choosing which wars to fight, and this seems likely to have little influence on how good they are at winning such wars or scaling weaponry.

Eg in the US often "which war" is much more political than "exactly what strategy should we use to win this war" is much more political than "how much fuel should our jets be able to carry", since more people can talk & speculate about the higher level questions. China's politics are much more closed than the US's, but you can bet similar dynamics are at play.

Replies from: weibac
comment by Milan W (weibac) · 2024-11-10T22:12:53.720Z · LW(p) · GW(p)

I should have been more clear. With "strategic ability", I was thinking about the kind of capabilities that let a government recognize which wars have good prospects, and to not initiate unfavorable wars despite ideological commitments.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-03-26T12:04:42.305Z · LW(p) · GW(p)

Novel Science is Inherently Illegible

Legibility, transparency, and open science are generally considered positive attributes, while opacity, elitism, and obscurantism are viewed as negative. However, increased legibility in science is not always beneficial and can often be detrimental.

Scientific management, with some exceptions, likely underperforms compared to simpler heuristics such as giving money to smart people or implementing grant lotteries. Scientific legibility suffers from the classic "Seeing like a State" problems. It constrains endeavors to the least informed stakeholder, hinders exploration, inevitably biases research to be simple and myopic, and exposes researchers to constant political tug-of-war between different interest groups poisoning objectivity. 

I think the above would be considered relatively uncontroversial in EA circles.  But I posit there is something deeper going on: 

Novel research is inherently illegible. If it were legible, someone else would have already pursued it. As science advances her concepts become increasingly counterintuitive and further from common sense. Most of the legible low-hanging fruit has already been picked, and novel research requires venturing higher into the tree, pursuing illegible paths with indirect and hard-to-foresee impacts.

Replies from: thomas-kwa, Seth Herd, ChristianKl, D0TheMath
comment by Thomas Kwa (thomas-kwa) · 2024-03-27T06:26:08.078Z · LW(p) · GW(p)

Novel research is inherently illegible.

I'm pretty skeptical of this and think we need data to back up such a claim. However there might be bias: when anyone makes a serendipitous discovery it's a better story, so it gets more attention. Has anyone gone through, say, the list of all Nobel laureates and looked at whether their research would have seemed promising before it produced results?

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-03-27T17:25:14.517Z · LW(p) · GW(p)

Thanks for your skepticism, Thomas. Before we get into this, I'd like to make sure actually disagree. My position is not that scientific progress is mostly due to plucky outsiders who are ignored for decades. (I feel something like this is a popular view on LW). Indeed, I think most scientific progress is made through pretty conventional (academic) routes.

I think one can predict that future scientific progress will likely be made by young smart people at prestigious universities and research labs specializing in fields that have good feedback loops and/or have historically made a lot of progress: physics, chemistry, medicine, etc

My contention is that beyond very broad predictive factors like this, judging whether a research direction is fruitful is hard & requires inside knowledge. Much of this knowledge is illegible, difficult to attain because it takes a lot of specialized knowledge etc.

Do you disagree with this ?

I do think that novel research is inherently illegible. Here are some thoughts on your comment :

1.Before getting into your Nobel prize proposal I'd like to caution for Hindsight bias (obvious reasons).

  1. And perhaps to some degree I'd like to argue the burden of proof should be on the converse: show me evidence that scientific progress is very legible. In some sense, predicting what directions will be fruitful is a bet against the (efficiënt ?) scientific market.

  2. I also agree the amount of prediction one can do will vary a lot. Indeed, it was itself an innovation (eg Thomas Edison and his lightbulbs !) that some kind of scientific and engineering progress could by systematized: the discovery of R&D.

I think this works much better for certain domains than for others and a to large degree the 'harder' & more 'novel' the problem is the more labs defer 'illegibly' to the inside knowledge of researchers.

Replies from: aysja, thomas-kwa
comment by aysja · 2024-03-29T09:35:07.084Z · LW(p) · GW(p)

I guess I'm not sure what you mean by "most scientific progress," and I'm missing some of the history here, but my sense is that importance-weighted science happens proportionally more outside of academia. E.g., Einstein did his miracle year outside of academia (and later stated that he wouldn't have been able to do it, had he succeeded at getting an academic position), Darwin figured out natural selection, and Carnot figured out the Carnot cycle, all mostly on their own, outside of academia. Those are three major scientists who arguably started entire fields (quantum mechanics, biology, and thermodynamics). I would anti-predict that future scientific progress, of the field-founding sort, comes primarily from people at prestigious universities, since they, imo, typically have some of the most intense gatekeeping dynamics which make it harder to have original thoughts. 

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-03-30T17:50:31.839Z · LW(p) · GW(p)

Good point. 

I do wonder to what degree that may be biased by the fact that there were vastly less academic positions before WWI/WWII. In the time of Darwin and Carnot these positions virtually didn't exist. In the time of Einstein they did exist but they were quite rare still. 

How many examples do you know of this happening past WWII?

Shannon was at Bell Labs iirc

As counterexample of field-founding happening in academia: Godel, Church, Turing were all in academia. 

comment by Thomas Kwa (thomas-kwa) · 2024-03-27T21:10:35.709Z · LW(p) · GW(p)

Oh, I actually 70% agree with this. I think there's an important distinction between legibility to laypeople vs legibility to other domain experts. Let me lay out my beliefs:

  • In the modern history of fields you mentioned, more than 70% of discoveries are made by people trying to discover the thing, rather than serendipitously.
  • Other experts in the field, if truth-seeking, are able to understand the theory of change behind the research direction without investing huge amounts of time.
  • In most fields, experts and superforecasters informed by expert commentary will have fairly strong beliefs about which approaches to a problem will succeed. The person working on something will usually have less than 1 bit advantage about whether their framework will be successful than the experts, unless they have private information (e.g. already did the crucial experiment). This is the weakest belief and I could probably be convinced otherwise just by anecdotes.
    • The successful researchers might be confident they will succeed, but unsuccessful ones could be almost as confident on average. So it's not that the research is illegible, it's just genuinely hard to predict who will succeed.
  • People often work on different approaches to the problem even if they can predict which ones will work. This could be due to irrationality, other incentives, diminishing returns to each approach, comparative advantage, etc.

If research were illegible to other domain experts, I think you would not really get Kuhnian paradigms, which I am pretty confident exist. Paradigm shifts mostly come from the track record of an approach, so maybe this doesn't count as researchers having an inside view of others' work though.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-03-28T01:04:20.078Z · LW(p) · GW(p)

Thank you, Thomas. I believe we find ourselves in broad agreement. The distinction you make between lay-legibility and expert-legibility is especially well-drawn.

One point: the confidence of researchers in their own approach may not be the right thing to look at. Perhaps a better measure is seeing who can predict not only their own approach will succed but explain in detail why other approaches won't work. Anecdotally, very succesful researchers have a keen sense of what will work out and what won't - in private conversation many are willing to share detailed models why other approaches will not work or are not as promising. I'd have to think about this more carefully but anecdotally the most succesful researchers have many bits of information over their competitors not just one or two. (Note that one bit of information means that their entire advantage could be wiped out by answering a single Y/N question. Not impossible, but not typical for most cases)

comment by Seth Herd · 2024-03-26T14:56:36.940Z · LW(p) · GW(p)

What areas of science are you thinking of? I think the discussion varies dramatically.

I think allowing less legibility would help make science less plodding, and allow it to move in larger steps. But there's also a question of what direction it's plodding. The problem I saw with psych and neurosci was that it tended to plod in nearly random, not very useful directions.

And what definition of "smart"? I'm afraid that by a common definition, smart people tend to do dumb research, in that they'll do galaxy brained projects that are interesting but unlikely to pay off. This is how you get new science, but not useful science.

In cognitive psychology and neuroscience, I want to see money given to people who are both creative and practical. They will do new science that is also useful.

In psychology and neuroscience, scientists pick the grantees, and they tend to give money to those whose research they understand. This produces an effect where research keeps following one direction that became popular long ago. I think a different method of granting would work better, but the particular method matters a lot.

Thinking about it a little more, having a mix of personality types involved would probably be useful. I always appreciated the contributions of the rare philospher who actually learned enough to join a discussion about psych or neurosci research.

I think the most important application of meta science theory is alignment research.

comment by ChristianKl · 2024-03-28T22:05:45.662Z · LW(p) · GW(p)

Novel research is inherently illegible. If it were legible, someone else would have already pursued it.

It might also be that a legible path would be low status to pursue in the existing scientific communities and thus nobody pursues it.

If you look at a low-hanging fruit that was unpicked for a long time, airborne transmission of many viruses like the common cold, is a good example. There's nothing illegible about it.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-03-30T17:51:34.220Z · LW(p) · GW(p)

mmm Good point. Do you have more examples?

Replies from: ChristianKl
comment by ChristianKl · 2024-03-31T14:36:24.679Z · LW(p) · GW(p)

The core reason for holding the belief is because the world does not look to me like there's little low hanging fruit in a variety of domains of knowledge I have thought about over the years. Of course it's generally not that easy to argue for the value of ideas that the mainstream does not care about publically.

Wei Dei recently wrote [LW · GW]:

I find it curious that none of my ideas have a following in academia or have been reinvented/rediscovered by academia (including the most influential ones so far UDT, UDASSA, b-money). Not really complaining, as they're already more popular than I had expected (Holden Karnofsky talked extensively about UDASSA on an 80,000 Hour podcast, which surprised me), it just seems strange that the popularity stops right at academia's door. 

If you look at the broader field of rationality, the work of Judea Pearl and that of Tetlock both could have been done twenty years earlier. Conceptually, I think you can argue that their work was some of the most important work that was done in the last decades.

Judea Pearl writes about how allergic people were against the idea of factoring in counterfactuals and causality. 

comment by Garrett Baker (D0TheMath) · 2024-03-26T15:49:31.808Z · LW(p) · GW(p)

I think the above would be considered relatively uncontroversial in EA circles.

I don’t think the application to EA itself would be uncontroversial.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-10-25T20:17:34.245Z · LW(p) · GW(p)

Why don't animals have guns? 

Or why didn't evolution evolve the Hydralisk?

Evolution has found (sometimes multiple times) the camera, general intelligence, nanotech, electronavigation, aerial endurance better than any drone, robots more flexible than any human-made drone, highly efficient photosynthesis, etc. 

First of all let's answer another question: why didn't evolution evolve the wheel like the alien wheeled elephants in His Dark Materials?

Is it biologically impossible to evolve?

Well, technically, the flagella of various bacteria is a proper wheel.

No the likely answer is that wheels are great when you have roads and suck when you don't. Roads are build by ants to some degree but on the whole probably don't make sense for an animal-intelligence species. 

Aren't there animals that use projectiles?

Hold up. Is it actually true that there is not a single animal with a gun, harpoon or other projectile weapon?

Porcupines have quils, some snakes spit venom, a type of fish spits water as a projectile to kick insects of leaves than eats insects. Bombadier beetles can produce an explosive chemical mixture. Skunks use some other chemicals. Some snails shoot harpoons from very close range. There is a crustacean that can snap its claw so quickly it creates a shockwave stunning fish. Octopi use ink.  Goliath birdeater spider shoot hair. Electric eels shoot electricity etc. 

Maybe there isn't an incentive gradient? The problem with this argument is that the same argument can be made for lots and lots of abilities that animals have developed, often multiple times. Flight, camera, a nervous system. 

But flight has an intermediate form: glider monkeys, flying squirrels, flying fish. 

Except, I think there are lots of intermediate forms for guns & harpoons too:

There are animals with quills. It's only a small number of steps from having quils that you release when attack to actively shooting and aiming these quils. Why didn't Evolution evolve Hydralisks? For many other examples - see the list above. 

In a Galaxy far far away

I think it is plausible that the reason animals don't have guns is simply an accident. Somewhere in the vast expanses of space circling a dim sun-like star the water-bearing planet Hiram Maxim is teeming with life. Nothing like an intelligent species has yet evolved yet it's many lifeforms sport a wide variety of highly effective projectile weapons. Indeed, the majority of larger lifeforms have some form of projective weapon as a result of the evolutionary arms race. The savannahs sport gazelle-like herbivores evading sniper-gun equppied predators. 

Some many parsecs away is the planet Big Bertha, a world is embroilled in permanent biological trench warfare. More than 95% percent of the biomass of animals larger than a mouse is taken up by members of just 4 geni of eusocial gun-equipped species or their domesticastes. Yet the individual intelligence of members of these species doesn't exceed that of a cat. 

The largest of the four geni builds massive dams like beavers, practices husbandry of various domesticated species, agriculture and engages in massive warfare against rival colonies using projectile harpoons that grow from their limbs. Yet all of this is biological, not technological: the behaviours and abilites are evolved rather than learned. There is not a single species whose intelligence rivals that of a Great ape, either individually or collectively. 

Replies from: dmurfet, nathan-helm-burger, D0TheMath, tao-lin, nim
comment by Daniel Murfet (dmurfet) · 2023-11-27T18:16:15.298Z · LW(p) · GW(p)

Please develop this question as a documentary special, for lapsed-Starcraft player homeschooling dads everywhere.

comment by Nathan Helm-Burger (nathan-helm-burger) · 2023-11-01T18:20:59.794Z · LW(p) · GW(p)

Most uses of projected venom or other unpleasant substance seem to be defensive rather than offensive. One reason for this is that it's expensive to make the dangerous substance, and throwing it away wastes it. This cost is affordable if it is used to save your own life, but not easily affordable to acquire a single meal. This life vs meal distinction plays into a lot of offense/defense strategy expenses.

For the hunting options, usually they are also useful for defense. The hunting options all seem cheaper to deploy: punching mantis shrimp, electric eel, fish spitting water...

My guess it that it's mostly a question of whether the intermediate steps to the evolved behavior are themselves advantageous. Having a path of consistently advantageous steps makes it much easier for something to evolve. Having to go through a trough of worse-in-the-short-term makes things much less likely to evolve. A projectile fired weakly is a cost (energy to fire, energy to producing firing mechanism, energy to produce the projectile, energy to maintain the complexity of the whole system despite it not being useful yet). Where's the payoff of a weakly fired projectile? Humans can jump that gap by intuiting that a faster projectile would be more effective. Evolution doesn't get to extrapolate and plan like that.

Replies from: carl-feynman, alexander-gietelink-oldenziel
comment by Carl Feynman (carl-feynman) · 2024-02-20T13:00:03.558Z · LW(p) · GW(p)

Jellyfish have nematocysts, which is a spear on a rope, with poison on the tip.  The spear has barbs, so when it goes in, it sticks.  Then the jellyfish pulls in its prey.  The spears are microscopic, but very abundant.

Replies from: nathan-helm-burger
comment by Nathan Helm-Burger (nathan-helm-burger) · 2024-02-21T04:19:02.457Z · LW(p) · GW(p)

Yes, but I think snake fangs and jellyfish nematocysts are a slightly different type of weapon. Much more targeted application of venom. If the jellyfish squirted their venom as a cloud into the water around them when a fish came near, I expect it would not be nearly as effective per unit of venom. As a case where both are present, the spitting cobra uses its fangs to inject venom into its prey. However, when threatened, it can instead (wastefully) spray out its venom towards the eyes of an attacker. (the venom has little effect on unbroken mammal skin, but can easily blind if it gets into their eyes).

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-11-01T18:49:48.078Z · LW(p) · GW(p)

Fair argument I guess where I'm lost is that I feel I can make the same 'no competitive intermediate forms' for all kinds of wondrous biological forms and functions that have evolved, e.g. the nervous system. Indeed, this kind of argument used to be a favorite for ID advocates.

Replies from: carl-feynman
comment by Carl Feynman (carl-feynman) · 2024-02-19T23:21:21.329Z · LW(p) · GW(p)

There are lots of excellent applications for even very simple nervous systems.  The simplest surviving nervous systems are those of jellyfish.  They form a ring of coupled oscillators around the periphery of the organism.  Their goal is to synchronize muscular contraction so the bell of the jellyfish contracts as one, to propel the jellyfish efficiently.  If the muscles contracted independently, it wouldn’t be nearly as good.

Any organism with eyes will profit from having a nervous system to connect the eyes to the muscles.  There’s a fungus with eyes and no nervous system, but as far as I know, every animal with eyes also has a nervous system. (The fungus in question is Pilobolus, which uses its eye to aim a gun.  No kidding!)

comment by Garrett Baker (D0TheMath) · 2023-10-25T21:30:11.890Z · LW(p) · GW(p)

My naive hypothesis: Once you're able to launch a projectile at a predator or prey such that it breaks skin or shell, if you want it to die, its vastly cheaper to make venom at the ends of the projectiles than to make the projectiles launch fast enough that there's a good increase in probability the adversary dies quickly.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-10-25T21:54:58.270Z · LW(p) · GW(p)

Why don't lions, tigers, wolves, crocodiles, etc have venom-tipped claws and teeth?

(Actually, apparently many ancestral mammal species like did have venom spurs, similar to the male platypus)

Replies from: JBlack
comment by JBlack · 2023-10-26T00:15:03.075Z · LW(p) · GW(p)

My completely naive guess would be that venom is mostly too slow for creatures of this size compared with gross physical damage and blood loss, and that getting close enough to set claws on the target is the hard part anyway. Venom seems more useful as a defensive or retributive mechanism than a hunting one.

comment by Tao Lin (tao-lin) · 2023-11-02T14:23:36.093Z · LW(p) · GW(p)

Another huge missed opportunity is thermal vision. Thermal infrared vision is a gigantic boon for hunting at night, and you might expect eg owls and hawks to use it to spot prey hundreds of meters away in pitch darkness, but no animals do (some have thermal sensing, but only extremely short range)

Replies from: carl-feynman, quetzal_rainbow, alexander-gietelink-oldenziel
comment by Carl Feynman (carl-feynman) · 2024-02-19T23:18:26.318Z · LW(p) · GW(p)

Snakes have thermal vision, using pits on their cheeks to form pinhole cameras. It pays to be cold-blooded when you’re looking for nice hot mice to eat.

comment by quetzal_rainbow · 2023-11-02T15:39:00.583Z · LW(p) · GW(p)

Thermal vision for warm-blooded animals has obvious problems with noise.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-11-03T08:11:28.036Z · LW(p) · GW(p)

Care to explain? Noise?

Replies from: quetzal_rainbow
comment by quetzal_rainbow · 2023-11-03T08:16:42.011Z · LW(p) · GW(p)

If you are warm, any warm-detectors inside your body will detect mostly you. Imagine if blood vessels in your own eye radiated in visible spectrum with the same intensity as daylight environment.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-11-03T17:19:44.090Z · LW(p) · GW(p)

Can't you filter that out? .

How do fighter planes do it?

Replies from: carl-feynman
comment by Carl Feynman (carl-feynman) · 2024-02-20T12:43:29.394Z · LW(p) · GW(p)

It‘s possible to filter out a constant high value, but not possible to filter out a high level of noise.  Unfortunately warmth = random vibration = noise.  If you want a low noise thermal camera, you have to cool the detector, or only look for hot things, like engine flares.  Fighter planes do both.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-11-03T08:10:55.139Z · LW(p) · GW(p)

Woah great example didn't know bout that. Thanks Tao

comment by nim · 2023-11-03T18:11:53.401Z · LW(p) · GW(p)

Animals do have guns. Humans are animals. Humans have guns. Evolution made us, we made guns, therefore guns indirectly exist because of evolution.

Or do you mean "why don't animals have something like guns but permanently attached to them instead of regular guns?" There, I'd start with wondering why humans prefer to have our guns separate from our bodies, compared to affixing them permanently or semi-permanently to ourselves. All the drawbacks of choosing a permanently attached gun would also disadvantage a hypothetical creature that got the accessory through a longer, slower selection process.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-20T17:39:53.318Z · LW(p) · GW(p)

Shower thought - why are sunglasses cool ?

Sunglasses create an asymmetry in the ability to discern emotions between the wearer and nonwearer. This implicitly makes the wearer less predictable, more mysterious, more dangerous and therefore higher in a dominance hierarchy.

Replies from: quila, NinaR, AllAmericanBreakfast, cubefox, skluug
comment by quila · 2024-10-20T18:27:46.671Z · LW(p) · GW(p)

also see ashiok from mtg: whole upper face/head is replaced with shadow

also, masks 'create an asymmetry in the ability to discern emotions' but do not seem to lead to the rest

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-20T22:51:10.306Z · LW(p) · GW(p)

That's a good counterexample ! Masks are dangerous and mysterious, but not cool in the way sunglasses are in agree

Replies from: D0TheMath, quila
comment by Garrett Baker (D0TheMath) · 2024-10-20T23:07:41.340Z · LW(p) · GW(p)

I think with sunglasses there’s a veneer of plausible deniability. They in fact have a utilitarian purpose outside of just creating information asymmetry. If you’re wearing a mask though, there’s no deniability. You just don’t want people to know where you’re looking.

Replies from: leogao
comment by leogao · 2024-10-21T06:08:14.476Z · LW(p) · GW(p)

there is an obvious utilitarian reason of not getting sick

Replies from: D0TheMath
comment by Garrett Baker (D0TheMath) · 2024-10-21T08:13:51.400Z · LW(p) · GW(p)

Oh I thought they meant like ski masks or something. For illness masks, the reason they’re not cool is very clearly that they imply you’re diseased.

(To a lesser extent too that your existing social status is so low you can’t expect to get away with accidentally infecting any friends or acquaintances, but my first point is more obvious & defensible)

comment by quila · 2024-10-21T08:03:09.707Z · LW(p) · GW(p)

oh i meant medical/covid ones. could also consider furry masks and the cat masks that femboys often wear (e.g. to obscure masculine facial structure), which feel cute rather than 'cool', though they are more like the natural human face in that they display an expression ("the face is a mask we wear over our skulls")

Replies from: D0TheMath
comment by Garrett Baker (D0TheMath) · 2024-10-21T08:14:25.425Z · LW(p) · GW(p)

Yeah pretty clearly these aren’t cool because they imply the wearer is diseased.

Replies from: quila
comment by quila · 2024-10-21T08:18:18.546Z · LW(p) · GW(p)

how? edit: maybe you meant just the first kind

Replies from: D0TheMath
comment by Garrett Baker (D0TheMath) · 2024-10-21T20:26:40.026Z · LW(p) · GW(p)

Yeah, I meant medical/covid masks imply the wearer is diseased. I would have also believed the cat mask is a medical/covid mask if you hadn't give a different reason for wearing it, so it has that going against it in terms of coolness. It also has a lack of plausible deniability going against it too. If you're wearing sunglasses there's actually a utilitarian reason behind wearing them outside of just creating information asymmetry. If you're just trying to obscure half your face, there's no such plausible deniability. You're just trying to obscure your face, so it becomes far less cool.

comment by Nina Panickssery (NinaR) · 2024-10-21T01:58:14.116Z · LW(p) · GW(p)

Isn't this already the commonly-accepted reason why sunglasses are cool?

Anyway, Claude agrees with you (see 1 and 3)

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-21T07:53:20.567Z · LW(p) · GW(p)

yes very lukewarm take

also nice product placement nina

comment by DirectedEvolution (AllAmericanBreakfast) · 2024-10-21T03:44:09.129Z · LW(p) · GW(p)

Sunglasses aren’t cool. They just tint the allure the wearer already has.

comment by cubefox · 2024-10-21T10:14:57.413Z · LW(p) · GW(p)

Follow-up question: If sunglasses are so cool, why do relatively few people wear them? Perhaps they aren't that cool after all?

Replies from: gwern
comment by gwern · 2024-10-22T01:23:25.701Z · LW(p) · GW(p)

Sunglasses can be too cool for most people to be able to wear in the absence of a good reason. Tom Cruise can go around wearing sun glasses any time he wants, and it'll look cool on him, because he's Tom Cruise. If we tried that, we would look like dorks because we're not cool enough to pull it off [LW · GW] and it would backfire on us. (Maybe our mothers would think we looked cool.) This could be said of many things: Tom Cruise or Kanye West or fashionable celebrities like them can go around wearing a fedora and trench coat and it'll look cool and he'll pull it off; but if anyone else tries it...

Replies from: cubefox
comment by cubefox · 2024-10-22T01:29:29.363Z · LW(p) · GW(p)

Yeah. I think the technical term for that would be cringe.

comment by Joey KL (skluug) · 2024-10-21T05:02:35.916Z · LW(p) · GW(p)

More reasons: people wear sunglasses when they’re doing fun things outdoors like going to the beach or vacationing so it’s associated with that, and also sometimes just hiding part of a picture can cause your brain to fill it in with a more attractive completion than is likely.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-13T20:37:14.249Z · LW(p) · GW(p)

My timelines are lengthening. 

I've long been a skeptic of scaling LLMs to AGI *. To me I fundamentally don't understand how this is even possible. It must be said that very smart people give this view credence. davidad, dmurfet. on the other side are vanessa kosoy and steven byrnes. When pushed proponents don't actually defend the position that a large enough transformer will create nanotech or even obsolete their job. They usually mumble something about scaffolding.

I won't get into this debate here but I do want to note that my timelines have lengthened, primarily because some of the never-clearly-stated but heavily implied AI developments by proponents of very short timelines have not materialized. To be clear, it has only been a year since gpt-4 is released, and gpt-5 is around the corner, so perhaps my hope is premature. Still my timelines are lengthening. 

A year ago, when gpt-3 came out progress was blindingly fast. Part of short timelines came from a sense of 'if we got surprised so hard by gpt2-3, we are completely uncalibrated, who knows what comes next'.

People seemed surprised by gpt-4 in a way that seemed uncalibrated to me. gpt-4 performance was basically in line with what one would expect if the scaling laws continued to hold. At the time it was already clear that the only really important driver was compute  data and that we would run out of both shortly after gpt-4. Scaling proponents suggested this was only the beginning, that there was a whole host of innovation that would be coming. Whispers of mesa-optimizers and simulators. 

One year in: Chain-of-thought doesn't actually improve things that much. External memory and super context lengths ditto. A whole list of proposed architectures seem to serve solely as a paper mill. Every month there is new hype about the latest LLM or image model. Yet they never deviate from expectations based on simple extrapolation of the scaling laws. There is only one thing that really seems to matter and that is compute and data. We have about 3 more OOMs of compute to go. Data may be milked another OOM. 

A big question will be whether gpt-5 will suddenly make agentGPT work ( and to what degree). It would seem that gpt-4 is in many ways far more capable than (most or all) humans yet agentGPT is curiously bad. 

All-in-all AI progress** is developing according to the naive extrapolations of Scaling Laws but nothing beyond that. The breathless twitter hype about new models is still there but it seems to be believed more at a simulacra level higher than I can parse. 

Does this mean we'll hit an AI winter? No. In my model there may be only one remaining roadblock to ASI (and I suspect I know what it is). That innovation could come at anytime. I don't know how hard it is, but I suspect it is not too hard. 

* the term AGI seems to denote vastly different things to different people in a way I find deeply confusing. I notice that the thing that I thought everybody meant by AGI is now being called ASI. So when I write AGI, feel free to substitute ASI. 

** or better, AI congress

addendum:  since I've been quoted in dmurfet's AXRP interview as believing that there are certain kinds of reasoning that cannot be represented by transformers/LLMs I want to be clear that this is not really an accurate portrayal of my beliefs. e.g. I don't think transformers don't truly understand, are just a stochastic parrot, or in other ways can't engage in the abstract reasoning that humans do. I think this is clearly false, as seen by interacting with any frontier model. 

Replies from: Vladimir_Nesov, Marcus Williams, faul_sname, adam-shai, zeshen, DanielFilan, dmurfet, stephen-mcaleese, stephen-mcaleese, dmurfet
comment by Vladimir_Nesov · 2024-05-13T21:54:08.207Z · LW(p) · GW(p)

With scale, there is visible improvement in difficulty of novel-to-chatbot ideas/details that is possible to explain in-context, things like issues with the code it's writing. If a chatbot is below some threshold of situational awareness of a task, no scaffolding can keep it on track, but for a better chatbot trivial scaffolding might suffice. Many people can't google for a solution to a technical issue, the difference between them and those who can is often subtle.

So modest amount of scaling alone seems plausibly sufficient for making chatbots that can do whole jobs almost autonomously. If this works, 1-2 OOMs more of scaling becomes both economically feasible and more likely to be worthwhile. LLMs think much faster, so they only need to be barely smart enough to help with clearing those remaining roadblocks.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-13T21:57:27.719Z · LW(p) · GW(p)

You may be right. I don't know of course. 

At this moment in time, it seems scaffolding tricks haven't really improved the baseline performance of models that much. Overwhelmingly, the capability comes down to whether the rlfhed base model can do the task.

Replies from: Vladimir_Nesov
comment by Vladimir_Nesov · 2024-05-13T22:30:12.817Z · LW(p) · GW(p)

it seems scaffolding tricks haven't really improved the baseline performance of models that much. Overwhelmingly, the capability comes down to whether the rlfhed base model can do the task.

That's what I'm also saying above (in case you are stating what you see as a point of disagreement). This is consistent with scaling-only short timeline expectations. The crux for this model is current chatbots being already close to autonomous agency and to becoming barely smart enough to help with AI research. Not them directly reaching superintelligence or having any more room for scaling.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-14T05:28:06.361Z · LW(p) · GW(p)

Yes agreed.

What I don't get about this position: If it was indeed just scaling - what's AI research for ? There is nothing to discover, just scale more compute. Sure you can maybe improve the speed of deploying compute a little but at the core of it it seems like a story that's in conflict with itself?

Replies from: nathan-helm-burger, Vladimir_Nesov
comment by Nathan Helm-Burger (nathan-helm-burger) · 2024-05-15T19:20:23.406Z · LW(p) · GW(p)

My view is that there's huge algorithmic gains in peak capability, training efficiency (less data, less compute), and inference efficiency waiting to be discovered, and available to be found by a large number of parallel research hours invested by a minimally competent multimodal LLM powered research team. So it's not that scaling leads to ASI directly, it's:

  1. scaling leads to brute forcing the LLM agent across the threshold of AI research usefulness
  2. Using these LLM agents in a large research project can lead to rapidly finding better ML algorithms and architectures.
  3. Training these newly discovered architectures at large scales leads to much more competent automated researchers.
  4. This process repeats quickly over a few months or years.
  5. This process results in AGI.
  6. AGI, if instructed (or allowed, if it's agentically motivated on its own to do so) to improve itself will find even better architectures and algorithms.
  7. This process can repeat until ASI. The resulting intelligence / capability / inference speed goes far beyond that of humans. 

Note that this process isn't inevitable, there are many points along the way where humans can (and should, in my opinion) intervene. We aren't disempowered until near the end of this.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-15T19:29:59.076Z · LW(p) · GW(p)

Why do you think there are these low-hanging algorithmic improvements?

Replies from: carl-feynman, nathan-helm-burger
comment by Carl Feynman (carl-feynman) · 2024-05-19T16:30:16.739Z · LW(p) · GW(p)

Here are two arguments for low-hanging algorithmic improvements.

First, in the past few years I have read many papers containing low-hanging algorithmic improvements.  Most such improvements are a few percent or tens of percent.  The largest such improvements are things like transformers or mixture of experts, which are substantial steps forward.  Such a trend is not guaranteed to persist, but that’s the way to bet.

Second, existing models are far less sample-efficient than humans.  We receive about a billion tokens growing to adulthood.  The leading LLMs get orders of magnitude more than that.  We should be able to do much better.  Of course, there’s no guarantee that such an improvement is “low hanging”.  

Replies from: Vladimir_Nesov
comment by Vladimir_Nesov · 2024-05-19T18:38:34.404Z · LW(p) · GW(p)

We receive about a billion tokens growing to adulthood. The leading LLMs get orders of magnitude more than that. We should be able to do much better.

Capturing this would probably be a big deal, but a counterpoint is that compute necessary to achieve an autonomous researcher using such sample efficient method might still be very large. Possibly so large that training an LLM with the same compute and current sample-inefficient methods is already sufficient to get a similarly effective autonomous researcher chatbot. In which case there is no effect on timelines. And given that the amount of data is not an imminent constraint on scaling [LW(p) · GW(p)], the possibility of this sample efficiency improvement being useless for the human-led stage of AI development won't be ruled out for some time yet.

Replies from: alexander-gietelink-oldenziel, carl-feynman
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-19T20:23:57.882Z · LW(p) · GW(p)

Could you train an LLM on pre 2014 Go games that could beat AlphaZero?

I rest my case.

Replies from: Vladimir_Nesov
comment by Vladimir_Nesov · 2024-05-19T20:57:49.465Z · LW(p) · GW(p)

The best method of improving sample efficiency might be more like AlphaZero. The simplest method that's more likely to be discovered might be more like training on the same data over and over with diminishing returns. Since we are talking low-hanging fruit, I think it's reasonable that first forays into significantly improved sample efficiency with respect to real data are not yet much better than simply using more unique real data.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-19T21:15:41.943Z · LW(p) · GW(p)

I would be genuinely surprised if training a transformer on the pre2014 human Go data over and over would lead it to spontaneously develop alphaZero capacity. I would expect it to do what it is trained to: emulate / predict as best as possible the distribution of human play. To some degree I would anticipate the transformer might develop some emergent ability that might make it slightly better than Go-Magnus - as we've seen in other cases - but I'd be surprised if this would be unbounded. This is simply not what the training signal is.

Replies from: Vladimir_Nesov
comment by Vladimir_Nesov · 2024-05-19T21:39:24.804Z · LW(p) · GW(p)

We start with an LLM trained on 50T tokens of real data, however capable it ends up being, and ask how to reach the same level of capability with synthetic data. If it takes more than 50T tokens of synthetic data, then it was less valuable per token than real data.

But at the same time, 500T tokens of synthetic data might train an LLM more capable than if trained on the 50T tokens of real data for 10 epochs. In that case, synthetic data helps with scaling capabilities beyond what real data enables, even though it's still less valuable per token.

With Go, we might just be running into the contingent fact of there not being enough real data to be worth talking about, compared with LLM data for general intelligence. If we run out of real data before some threshold of usefulness, synthetic data becomes crucial (which is the case with Go). It's unclear if this is the case for general intelligence with LLMs, but if it is, then there won't be enough compute to improve the situation unless synthetic data also becomes better per token, and not merely mitigates the data bottleneck and enables further improvement given unbounded compute.

I would be genuinely surprised if training a transformer on the pre2014 human Go data over and over would lead it to spontaneously develop alphaZero capacity.

I expect that if we could magically sample much more pre-2014 unique human Go data than was actually generated by actual humans (rather than repeating the limited data we have), from the same platonic source and without changing the level of play, then it would be possible to cheaply tune an LLM trained on it to play superhuman Go.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-20T07:39:47.406Z · LW(p) · GW(p)

I don't know what you mean by 'general intelligence' exactly but I suspect you mean something like human+ capability in a broad range of domains. I agree LLMs will become generally intelligent in this sense when scaled, arguably even are, for domains with sufficient data. But that's kind of the sticker right? Cave men didn't have the whole internet to learn from yet somehow did something that not even you seem to claim LLMs will be able to do: create the (date of the) Internet.

(Your last claim seems surprising. Pre-2014 games don't have close to the ELO of alphaZero. So a next-token would be trained to simulate a human player up tot 2800, not 3200+. )

Replies from: Vladimir_Nesov
comment by Vladimir_Nesov · 2024-05-21T18:17:28.011Z · LW(p) · GW(p)

Pre-2014 games don't have close to the ELO of alphaZero. So a next-token would be trained to simulate a human player up to 2800, not 3200+.

Models can be thought of as repositories of features rather than token predictors. A single human player knows some things, but a sufficiently trained model knows all the things that any of the players know. Appropriately tuned, a model might be able to tap into this collective knowledge to a greater degree than any single human player. Once the features are known, tuning and in-context learning that elicit their use are very sample efficient.

This framing seems crucial for expecting LLMs to reach researcher level of capability given a realistic amount of data, since most humans are not researchers, and don't all specialize in the same problem. The things researcher LLMs would need to succeed in learning are cognitive skills, so that in-context performance gets very good at responding to novel engineering and research agendas only seen in-context (or a certain easier feat that I won't explicitly elaborate on).

Cave men didn't have the whole internet to learn from yet somehow did something that not even you seem to claim LLMs will be able to do: create the (date of the) Internet.

Possibly the explanation for the Sapient Paradox, that prehistoric humans managed to spend on the order of 100,000 years without developing civilization, is that they lacked cultural knowledge of crucial general cognitive skills. Sample efficiency of the brain enabled their fixation in language across cultures and generations, once they were eventually distilled, but it took quite a lot of time.

Modern humans and LLMs start with all these skills already available in the data, though humans can more easily learn them. LLMs tuned to tap into all of these skills at the same time might be able to go a long way without an urgent need to distill new ones, merely iterating on novel engineering and scientific challenges, applying the same general cognitive skills over and over.

comment by Carl Feynman (carl-feynman) · 2024-05-19T19:38:13.328Z · LW(p) · GW(p)

When I brought up sample inefficiency, I was supporting Mr. Helm-Burger‘s statement that “there's huge algorithmic gains in …training efficiency (less data, less compute) … waiting to be discovered”.  You’re right of course that a reduction in training data will not necessarily reduce the amount of computation needed.  But once again, that’s the way to bet.

Replies from: Vladimir_Nesov
comment by Vladimir_Nesov · 2024-05-19T20:52:24.257Z · LW(p) · GW(p)

a reduction in training data will not necessarily reduce the amount of computation needed. But once again, that’s the way to bet

I'm ambivalent on this. If the analogy between improvement of sample efficiency and generation of synthetic data holds, synthetic data seems reasonably likely to be less valuable than real data (per token). In that case we'd be using all the real data we have anyway, which with repetition is sufficient for up to about $100 billion training runs (we are at $100 million right now). Without autonomous agency (not necessarily at researcher level) before that point, there won't be investment to go over that scale until much later, when hardware improves and the cost goes down.

comment by Nathan Helm-Burger (nathan-helm-burger) · 2024-05-16T01:53:46.195Z · LW(p) · GW(p)

My answer to that is currently in the form of a detailed 2 hour lecture with a bibliography that has dozens of academic papers in it, which I only present to people that I'm quite confident aren't going to spread the details. It's a hard thing to discuss in detail without sharing capabilities thoughts. If I don't give details or cite sources, then... it's just, like, my opinion, man. So my unsupported opinion is all I have to offer publicly. If you'd like to bet on it, I'm open to showing my confidence in my opinion by betting that the world turns out how I expect it to.

comment by Vladimir_Nesov · 2024-05-14T14:51:08.262Z · LW(p) · GW(p)

a story that's in conflict with itself

The story involves phase changes. Just scaling is what's likely to be available to human developers in the short term (a few years), it's not enough for superintelligence. Autonomous agency secures funding for a bit more scaling. If this proves sufficient to get smart autonomous chatbots, they then provide speed to very quickly reach the more elusive AI research needed for superintelligence.

It's not a little speed, it's a lot of speed, serial speedup of about 100x plus running in parallel. This is not as visible today, because current chatbots are not capable of doing useful work with serial depth, so the serial speedup is not in practice distinct from throughput and cost. But with actually useful chatbots it turns decades to years, software and theory from distant future become quickly available, non-software projects get to be designed in perfect detail faster than they can be assembled.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-18T20:51:59.865Z · LW(p) · GW(p)

In my mainline model there are only a few innovations needed, perhaps only a single big one to product an AGI which just like the Turing Machine sits at the top of the Chomsky Hierarchy will be basically the optimal architecture given resource constraints. There are probably some minor improvements todo with bridging the gap between theoretically optimal architecture and the actual architecture, or parts of the algorithm that can be indefinitely improved but with diminishing returns (these probably exist due to Levin and possibly.matrix.multiplication is one of these). On the whole I expect AI research to be very chunky.

Indeed, we've seen that there was really just one big idea to all current AI progress: scaling, specifically scaling GPUs on maximally large undifferentiated datasets. There were some minor technical innovations needed to pull this off but on the whole that was the clinger.

Of course, I don't know. Nobody knows. But I find this the most plausible guess based on what we know about intelligence, learning, theoretical computer science and science in general.

Replies from: Vladimir_Nesov, Vladimir_Nesov
comment by Vladimir_Nesov · 2024-05-19T13:31:39.704Z · LW(p) · GW(p)

(Re: Difficult to Parse react on the other comment [LW(p) · GW(p)]
I was confused about relevance of your comment above [LW(p) · GW(p)] on chunky innovations, and it seems to be making some point (for which what it actually says is an argument), but I can't figure out what it is. One clue was that it seems like you might be talking about innovations needed for superintelligence, while I was previously talking about possible absence of need for further innovations to reach autonomous researcher chatbots, an easier target. So I replied with formulating this distinction and some thoughts on the impact and conditions for reaching innovations of both kinds. Possibly the relevance of this was confusing in turn.)

comment by Vladimir_Nesov · 2024-05-18T21:48:01.237Z · LW(p) · GW(p)

There are two kinds of relevant hypothetical innovations: those that enable chatbot-led autonomous research, and those that enable superintelligence. It's plausible that there is no need for (more of) the former, so that mere scaling through human efforts will lead to such chatbots in a few years regardless. (I think it's essentially inevitable that there is currently enough compute that with appropriate innovations we can get such autonomous human-scale-genius chatbots, but it's unclear if these innovations are necessary or easy to discover.) If autonomous chatbots are still anything like current LLMs, they are very fast compared to humans, so they quickly discover remaining major innovations of both kinds.

In principle, even if innovations that enable superintelligence (at scale feasible with human efforts in a few years) don't exist at all, extremely fast autonomous research and engineering still lead to superintelligence, because they greatly accelerate scaling. Physical infrastructure might start scaling really fast using pathways like macroscopic biotech even if drexlerian nanotech is too hard without superintelligence or impossible in principle. Drosophila biomass doubles every 2 days, small things can assemble into large things.

comment by Marcus Williams · 2024-05-13T21:28:03.833Z · LW(p) · GW(p)

Wasn't the surprising thing about GPT-4 that scaling laws did hold? Before this many people expected scaling laws to stop before such a high level of capabilities. It doesn't seem that crazy to think that a few more OOMs could be enough for greater than human intelligence. I'm not sure that many people predicted that we would have much faster than scaling law progress (at least until ~human intelligence AI can speed up research)? I think scaling laws are the extreme rate of progress which many people with short timelines worry about.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-13T21:45:18.580Z · LW(p) · GW(p)

To some degree yes, they were not guaranteed to hold. But by that point they held for over 10 OOMs iirc and there was no known reason they couldn't continue.

This might be the particular twitter bubble I was in but people definitely predicted capabilities beyond simple extrapolation of scaling laws.

comment by faul_sname · 2024-05-14T00:52:16.951Z · LW(p) · GW(p)

When pushed proponents don't actually defend the position that a large enough transformer will create nanotech

Can you expand on what you mean by "create nanotech?" If improvements to our current photolithography techniques count, I would not be surprised if (scaffolded) LLMs could be useful for that. Likewise for getting bacteria to express polypeptide catalysts for useful reactions, and even maybe figure out how to chain several novel catalysts together to produce something useful (again, referring to scaffolded LLMs with access to tools).

If you mean that LLMs won't be able to bootstrap from our current "nanotech only exists in biological systems and chip fabs" world to Drexler-style nanofactories, I agree with that, but I expect things will get crazy enough that I can't predict them long before nanofactories are a thing (if they ever are).

or even obsolete their job

Likewise, I don't think LLMs can immediately obsolete all of the parts of my job. But they sure do make parts of my job a lot easier. If you have 100 workers that each spend 90% of their time on one specific task, and you automate that task, that's approximately as useful as fully automating the jobs of 90 workers. "Human-equivalent" is one of those really leaky abstractions -- I would be pretty surprised if the world had any significant resemblance to the world of today by the time robotic systems approached the dexterity and sensitivity of human hands for all of the tasks we use our hands for, whereas for the task of "lift heavy stuff" or "go really fast" machines left us in the dust long ago.

Iterative improvements on the timescale we're likely to see are still likely to be pretty crazy by historical standards. But yeah, if your timelines were "end of the world by 2026" I can see why they'd be lengthening now.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-18T20:40:43.434Z · LW(p) · GW(p)

My timelines were not 2026. In fact, I made bets against doomers 2-3 years ago, one will resolve by next year.

I agree iterative improvements are significant. This falls under "naive extrapolation of scaling laws".

By nanotech I mean something akin to drexlerian nanotech or something similarly transformative in the vicinity. I think it is plausible that a true ASI will be able to make rapid progress (perhaps on the order of a few years or a decade) on nanotech. I suspect that people that don't take this as a serious possibility haven't really thought through what AGI/ASI means + what the limits and drivers of science and tech really are; I suspect they are simply falling prey to status-quo bias.

comment by Adam Shai (adam-shai) · 2024-05-14T01:27:44.636Z · LW(p) · GW(p)

Lengthening from what to what?

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-14T05:30:48.980Z · LW(p) · GW(p)

I've never done explicit timelines estimates before so nothing to compare to. But since it's a gut feeling anyway, I'm saying my gut is lengthening.

comment by zeshen · 2024-05-14T10:38:23.464Z · LW(p) · GW(p)

Agreed [LW(p) · GW(p)]. I'm also pleasantly surprised that your take isn't heavily downvoted.

comment by DanielFilan · 2024-05-14T22:46:30.932Z · LW(p) · GW(p)

Links to Dan Murfet's AXRP interview:

comment by Daniel Murfet (dmurfet) · 2024-05-14T10:05:03.545Z · LW(p) · GW(p)

I don't recall what I said in the interview about your beliefs, but what I meant to say was something like what you just said in this post, apologies for missing the mark.

comment by Stephen McAleese (stephen-mcaleese) · 2024-05-15T18:45:22.499Z · LW(p) · GW(p)

State-of-the-art models such as Gemini aren't LLMs anymore. They are natively multimodal or omni-modal transformer models that can process text, images, speech and video. These models seem to me like a huge jump in capabilities over text-only LLMs like GPT-3.

comment by Stephen McAleese (stephen-mcaleese) · 2024-05-14T08:29:46.613Z · LW(p) · GW(p)

Chain-of-thought prompting makes models much more capable. In the original paper "Chain-of-Thought Prompting Elicits Reasoning in Large Language Models", PaLM 540B with standard prompting only solves 18% of problems but 57% of problems with chain-of-thought prompting.

I expect the use of agent features such as reflection will lead to similar large increases in capabilities as well in the near future.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-14T08:32:04.306Z · LW(p) · GW(p)

Those numbers don't really accord with my experience actually using gpt-4. Generic prompting techniques just don't help all that much.

Replies from: stephen-mcaleese
comment by Stephen McAleese (stephen-mcaleese) · 2024-05-14T10:26:32.495Z · LW(p) · GW(p)

I just asked GPT-4 a GSM8K problem and I agree with your point. I think what's happening is that GPT-4 has been fine-tuned to respond with chain-of-thought reasoning by default so it's no longer necessary to explicitly ask it to reason step-by-step. Though if you ask it to "respond with just a single number" to eliminate the chain-of-thought reasoning it's problem-solving ability is much worse.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-04-10T19:09:28.215Z · LW(p) · GW(p)

Encrypted Batteries 

(I thank Dmitry Vaintrob for the idea of encrypted batteries. Thanks to Adam Scholl for the alignment angle. Thanks to the Computational Mechanics at the receent compMech conference. )

There are no Atoms in the Void just Bits in the Description. Given the right string a Maxwell Demon transducer can extract energy from a heatbath. 

Imagine a pseudorandom heatbath + nano-Demon. It looks like a heatbath from the outside but secretly there is a private key string that when fed to the nano-Demon allows it to extra lots of energy from the heatbath. 

 

P.S. Beyond the current ken of humanity lies a generalized concept of free energy that describes the generic potential ability or power of an agent to achieve goals. Money, the golden calf of Baal is one of its many avatars. Could there be ways to encrypt generalized free energy batteries to constraint the user to only see this power for good? It would be like money that could be only spent on good things. 

Replies from: gwern
comment by gwern · 2024-04-11T01:32:29.068Z · LW(p) · GW(p)

Imagine a pseudorandom heatbath + nano-Demon. It looks like a heatbath from the outside but secretly there is a private key string that when fed to the nano-Demon allows it to extra lots of energy from the heatbath.

What would a 'pseudorandom heatbath' look like? I would expect most objects to quickly depart from any sort of private key or PRNG. Would this be something like... a reversible computer which shuffles around a large number of blank bits in a complicated pseudo-random order every timestep*, exposing a fraction of them to external access? so a daemon with the key/PRNG seed can write to the blank bits with approaching 100% efficiency (rendering it useful for another reversible computer doing some actual work) but anyone else can't do better than 50-50 (without breaking the PRNG/crypto) and that preserves the blank bit count and is no gain?

* As I understand reversible computing, you can have a reversible computer which does that for free: if this is something like a very large period loop blindly shuffling its bits, it need erase/write no bits (because it's just looping through the same states forever, akin to a time crystal), and so can be computed indefinitely at arbitrarily low energy cost. So any external computer which syncs up to it can also sync at zero cost, and just treat the exposed unused bits as if they were its own, thereby saving power.

Replies from: alexander-gietelink-oldenziel, MakoYass
comment by mako yass (MakoYass) · 2024-04-11T18:46:16.615Z · LW(p) · GW(p)

Yeah I'm pretty sure you would need to violate heisenberg uncertainty in order to make this and then you'd have to keep it in a 0 kelvin cleanroom forever.

A practical locked battery with tamperproofing would mostly just look like a battery.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-23T12:10:26.733Z · LW(p) · GW(p)

AGI companies merging within next 2-3 years inevitable?

There are currently about a dozen major AI companies racing towards AGI with many more minor AI companies. The way the technology shakes out this seems like unstable equilibrium. 

It seems by now inevitable that we will see further mergers, joint ventures - within two years there might only be two or three major players left. Scale is all-dominant. There is no magic sauce, no moat. OpenAI doesn't have algorithms that her competitors can't copy within  6-12 months. It's all leveraging compute. Whatever innovations smaller companies make can be easily stolen by tech giants. 

e.g. we might have xAI- Meta, Anthropic- DeepMind-SSI-Google, OpenAI-Microsoft-Apple. 

Actuallly, although this would be deeply unpopular in EA circles it wouldn't be all that surprising if Anthropic and OpenAI would team up. 

And - of course - a few years later we might only have two competitors: USA, China. 

EDIT: the obvious thing to happen is that nvidia realizes it can just build AI itself. if Taiwan is Dune, GPUs are the spice, then nvidia is house Atreides

Replies from: bogdan-ionut-cirstea, leon-lang, Vladimir_Nesov, Mo Nastri, bogdan-ionut-cirstea
comment by Bogdan Ionut Cirstea (bogdan-ionut-cirstea) · 2024-10-23T13:52:26.907Z · LW(p) · GW(p)

Whatever innovations smaller companies make can be easily stolen by tech giants. 

And they / their basic components are probably also published by academia, though the precise hyperparameters, etc. might still matter and be non-trivial/costly to find.

comment by Leon Lang (leon-lang) · 2024-10-23T12:16:47.765Z · LW(p) · GW(p)

I have a similar feeling, but there are some forces in the opposite direction:

  • Nvidia seems to limit how many GPUs a single competitor can acquire.
  • training frontier models becomes cheaper over time. Thus, those that build competitive models some time later than the absolute frontier have to invest much less resources.
comment by Vladimir_Nesov · 2024-10-23T14:25:36.059Z · LW(p) · GW(p)

In 2-3 years they would need to decide on training systems built in 3-5 years, and by 2027-2029 the scale might get to $200-1000 billion [LW(p) · GW(p)] for an individual training system. (This is assuming geographically distributed training is solved, since such systems would need 5-35 gigawatts.)

Getting to a go-ahead on $200 billion systems might require a level of success that also makes $1 trillion plausible. So instead of merging, they might instead either temporarily give up on scaling further (if there isn't sufficient success in 2-3 years), or become capable of financing such training systems individually, without pooling efforts.

comment by Mo Putera (Mo Nastri) · 2024-10-23T15:09:08.651Z · LW(p) · GW(p)

the obvious thing to happen is that nvidia realizes it can just build AI itself. if Taiwan is Dune, GPUs are the spice, then nvidia is house Atreides

They've already started... 

comment by Bogdan Ionut Cirstea (bogdan-ionut-cirstea) · 2024-10-23T14:23:01.119Z · LW(p) · GW(p)

For similar arguments, I think it's gonna be very hard/unlikely to stop China from having AGI within a couple of years of the US (and most relevant AI chips currently being produced in Taiwan should probably further increase the probability of this). So taking on a lot more x-risk to try and race hard vs. China doesn't seem like a good strategy from this POV.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-07-05T12:55:21.012Z · LW(p) · GW(p)

Current work on Markov blankets and Boundaries on LW is flawed and outdated. State of the art should factor through this paper on Causal Blankets; https://iwaiworkshop.github.io/papers/2020/IWAI_2020_paper_22.pdf

A key problem for accounts of blankets and boundaries I have seen on LW so far is the following elementary problem (from the paper):
"Therefore, the MB [Markov Blanket] formalism forbids interdependencies induced by past events that are kept in memory, but may not directly influence the present state of the blankets.

Thanks to Fernando Rosas telling me about this paper. 

Replies from: Gunnar_Zarncke, LosPolloFowler
comment by Gunnar_Zarncke · 2024-07-05T16:15:19.758Z · LW(p) · GW(p)

You may want to make this a linkpost to that paper as that can then be tagged and may be noticed more widely.

comment by Stephen Fowler (LosPolloFowler) · 2024-07-06T04:37:12.298Z · LW(p) · GW(p)

I have only skimmed the paper.

Is my intuition correct that in the MB formalism, past events that are causally linked to are not included in the Markov Blanket, but the node corresponding to the memory state still is included in the MB?

That is, the influence of the past event is mediated by a node corresponding to having memory of that past event?

Replies from: mateusz-baginski
comment by Mateusz Bagiński (mateusz-baginski) · 2024-07-30T10:36:06.442Z · LW(p) · GW(p)

Well, past events--before some time t--kind of obviously can't be included in the Markov blanket at time t.

As far as I understand it, the MB formalism captures only momentary causal interactions between "Inside" and "Outside" but doesn't capture a kind of synchronicity/fine-tuning-ish statistical dependency that doesn't manifest in the current causal interactions (across the Markov blanket) but is caused by past interactions.

For example, if you learned a perfect weather forecast for the next month and then went into a completely isolated bunker but kept track of what day it was, your beliefs and the actual weather would be very dependent even though there's no causal interaction (after you entered the bunker) between your beliefs and the weather. This is therefore omitted by MBs and CBs want to capture that.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-14T11:22:49.005Z · LW(p) · GW(p)

Problem of Old Evidence, the Paradox of Ignorance and Shapley Values

Paradox of Ignorance

Paul Christiano presents the "paradox of ignorance" where a weaker, less informed agent appears to outperform a more powerful, more informed agent in certain situations. This seems to contradict the intuitive desideratum that more information should always lead to better performance.

The example given is of two agents, one powerful and one limited, trying to determine the truth of a universal statement ∀x:ϕ(x) for some Δ0 formula ϕ. The limited agent treats each new value of ϕ(x) as a surprise and evidence about the generalization ∀x:ϕ(x). So it can query the environment about some simple inputs x and get a reasonable view of the universal generalization.

In contrast, the more powerful agent may be able to deduce ϕ(x) directly for simple x. Because it assigns these statements prior probability 1, they don't act as evidence at all about the universal generalization ∀x:ϕ(x). So the powerful agent must consult the environment about more complex examples and pay a higher cost to form reasonable beliefs about the generalization.

Is it really a problem?

However, I argue that the more powerful agent is actually justified in assigning less credence to the universal statement ∀x:ϕ(x). The reason is that the probability mass provided by examples x₁, ..., xₙ such that ϕ(xᵢ) holds is now distributed among the universal statement ∀x:ϕ(x) and additional causes Cⱼ known to the more powerful agent that also imply ϕ(xᵢ). Consequently, ∀x:ϕ(x) becomes less "necessary" and has less relative explanatory power for the more informed agent.

An implication of this perspective is that if the weaker agent learns about the additional causes Cⱼ, it should also lower its credence in ∀x:ϕ(x).

More generally, we would like the credence assigned to propositions P (such as ∀x:ϕ(x)) to be independent of the order in which we acquire new facts (like xᵢ, ϕ(xᵢ), and causes Cⱼ).

Shapley Value

The Shapley value addresses this limitation by providing a way to average over all possible orders of learning new facts. It measures the marginal contribution of an item (like a piece of evidence) to the value of sets containing that item, considering all possible permutations of the items. By using the Shapley value, we can obtain an order-independent measure of the contribution of each new fact to our beliefs about propositions like ∀x:ϕ(x).

Further thoughts

I believe this is closely related, perhaps identical, to the 'Problem of Old Evidence' [? · GW] as considered by Abram Demski.

Suppose a new scientific hypothesis, such as general relativity, explains a well-know observation such as the perihelion precession of mercury better than any existing theory. Intuitively, this is a point in favor of the new theory. However, the probability for the well-known observation was already at 100%. How can a previously-known statement provide new support for the hypothesis, as if we are re-updating on evidence we've already updated on long ago? This is known as the problem of old evidence, and is usually levelled as a charge against Bayesian epistemology.

 

[Thanks to @Jeremy Gillen [LW · GW] for pointing me towards this interesting Christiano paper]

Replies from: abramdemski, jeremy-gillen, kromem, cubefox
comment by abramdemski · 2024-05-28T01:04:45.717Z · LW(p) · GW(p)

The matter seems terribly complex and interesting to me.

Notions of Accuracy?

Suppose  is a prior which has uncertainty about  and uncertainty about . This is the more ignorant prior. Consider  some prior which has the same beliefs about the universal statement --  -- but which knows  and .

We observe that  can increase its credence in the universal statement by observing the first two instances,  and , while  cannot do this --  needs to wait for further evidence. This is interpreted as a defect.

The moral is apparently that a less ignorant prior can be worse than a more ignorant one; more specifically, it can learn more slowly.

However, I think we need to be careful about the informal notion of "more ignorant" at play here. We can formalize this by imagining a numerical measure of the accuracy of a prior. We might want it to be the case that more accurate priors are always better to start with. Put more precisely: a more accurate prior should also imply a more accurate posterior after updating. Paul's example challenges this notion, but he does not prove that no plausible notion of accuracy will have this property; he only relies on an informal notion of ignorance.

So I think the question is open: when can a notion of accuracy fail to follow the rule "more accurate priors yield more accurate posteriors"? EG, can a proper scoring rule fail to meet this criterion? This question might be pretty easy to investigate.

Conditional probabilities also change?

I think the example rests on an intuitive notion that we can construct  by imagining  but modifying it to know  and . However, the most obvious way to modify it so is by updating on those sentences. This fails to meet the conditions of the example, however;  would already have an increased probability for the universal statement.

So, in order to move the probability of  and  upwards to 1 without also increasing the probability of the universal, we must do some damage to the probabilistic relationship between the instances and the universal. The prior  doesn't just know   and ; it also believes the conditional probability of the universal statement given those two sentences to be lower than  believes them to be.

It doesn't think it should learn from them!

This supports Alexander's argument that there is no paradox, I think. However, I am not ultimately convinced. Perhaps I will find more time to write about the matter later.

Replies from: abramdemski
comment by abramdemski · 2024-05-28T14:09:50.886Z · LW(p) · GW(p)

(continued..)

Explanations?

Alexander analyzes the difference between  and  in terms of the famous "explaining away" effect. Alexander supposes that  has learned some "causes":

The reason is that the probability mass provided by examples x₁, ..., xₙ such that ϕ(xᵢ) holds is now distributed among the universal statement ∀x:ϕ(x) and additional causes Cⱼ known to the more powerful agent that also imply ϕ(xᵢ). Consequently, ∀x:ϕ(x) becomes less "necessary" and has less relative explanatory power for the more informed agent.

An implication of this perspective is that if the weaker agent learns about the additional causes Cⱼ, it should also lower its credence in ∀x:ϕ(x).

Postulating these causes adds something to the scenario. One possible view is that Alexander is correct so far as Alexander's argument goes, but incorrect if there are no such  to consider.

However, I do not find myself endorsing Alexander's argument even that far.

If  and  have a common form, or are correlated in some way -- so there is an explanation which tells us why the first two sentences,  and , are true, and which does not apply to  -- then I agree with Alexander's argument.

If  and  are uncorrelated, then it starts to look like a coincidence. If I find a similarly uncorrelated  for  for , and a few more, then it will feel positively unexplained. Although each explanation is individually satisfying, nowhere do I have an explanation of why all of them are turning up true.

I think the probability of the universal sentence should go up at this point.

So, what about my "conditional probabilities also change" variant of Alexander's argument? We might intuitively think that  and  should be evidence for the universal generalization, but  does not believe this -- its conditional probabilities indicate otherwise. 

I find this ultimately unconvincing because the point of Paul's example, in my view, is that more accurate priors do not imply more accurate posteriors. I still want to understand what conditions can lead to this (including whether it is true for all notions of "accuracy" satisfying some reasonable assumptions EG proper scoring rules).

Another reason I find it unconvincing is because even if we accepted this answer for the paradox of ignorance, I think it is not at all convincing for the problem of old evidence. 

What is the 'problem' in the problem of old evidence?

... to be further expanded later ...

comment by Jeremy Gillen (jeremy-gillen) · 2024-05-14T11:52:52.283Z · LW(p) · GW(p)

This doesn't feel like it resolves that confusion for me, I think it's still a problem with the agents he describes in that paper.

The causes  are just the direct computation of  for small values of . If they were arguments that only had bearing on small values of x and implied nothing about larger values (e.g. an adversary selected some  to show you, but filtered for  such that ), then it makes sense that this evidence has no bearing on. But when there was no selection or other reason that the argument only applies to small , then to me it feels like the existence of the evidence (even though already proven/computed) should still increase the credence of the forall.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-14T14:21:57.864Z · LW(p) · GW(p)

I didn't intend the causes to equate to direct computation of \phi(x) on the x_i. They are rather other pieces of evidence that the powerful agent has that make it believe \phi(x_i). I don't know if that's what you meant.

I agree seeing x_i such that \phi(x_i) should increase credence in \forall x \phi(x) even in the presence of knowledge of C_j. And the Shapely value proposal will do so.

(Bad tex. On my phone)

comment by kromem · 2024-05-15T01:25:43.146Z · LW(p) · GW(p)

It's funny that this has been recently shown in a paper. I've been thinking a lot about this phenomenon regarding fields with little to no capacity for testable predictions like history.

I got very into history over the last few years, and found there was a significant advantage to being unknowledgeable that was not available to the knowledged, and it was exactly what this paper is talking about.

By not knowing anything, I could entertain multiple bizarre ideas without immediately thinking "but no, that doesn't make sense because of X." And then, each of those ideas becomes in effect its own testable prediction. If there's something to it, as I learn more about the topic I'm going to see significantly more samples of indications it could be true and few convincing to the contrary. But if it probably isn't accurate, I'll see few supporting samples and likely a number of counterfactual examples.

You kind of get to throw everything at the wall and see what sticks over time.

In particular, I found that it was especially powerful at identifying clustering trends in cross-discipline emerging research in things that were testable, such as archeological finds and DNA results, all within just the past decade, which despite being relevant to the field of textual history is still largely ignored in the face of consensus built on conviction.

It reminds me a lot of science historian John Helibron's quote, "The myth you slay today may contain a truth you need tomorrow."

If you haven't had the chance to slay any myths, you also haven't preemptively killed off any truths along with it.

Replies from: gwern
comment by gwern · 2024-05-15T18:43:37.751Z · LW(p) · GW(p)

One of the interesting thing about AI minds (such as LLMs) is that in theory, you can turn many topics into testable science while avoiding the 'problem of old evidence', because you can now construct artificial minds and mold them like putty. They know what you want them to know, and so you can see what they would predict in the absence of knowledge, or you can install in them false beliefs to test out counterfactual intellectual histories, or you can expose them to real evidence in different orders to measure biases or path dependency in reasoning.

With humans, you can't do that because they are so uncontrolled: even if someone says they didn't know about crucial piece of evidence X, there is no way for them to prove that, and they may be honestly mistaken and have already read about X and forgotten it (but humans never really forget so X has already changed their "priors", leading to double-counting), or there is leakage. And you can't get people to really believe things at the drop of a hat, so you can't make people imagine, "suppose Napoleon had won Waterloo, how do you predict history would have changed?" because no matter how you try to participate in the spirit of the exercise, you always know that Napoleon lost and you have various opinions on that contaminating your retrodictions, and even if you have never read a single book or paper on Napoleon, you are still contaminated by expressions like "his Waterloo" ('Hm, the general in this imaginary story is going to fight at someplace called Waterloo? Bad vibes. I think he's gonna lose.')

But with a LLM, say, you could simply train it with all timestamped texts up to Waterloo, like all surviving newspapers, and then simply have one version generate a bunch of texts about how 'Napoleon won Waterloo', train the other version on these definitely-totally-real French newspaper reports about his stunning victory over the monarchist invaders, and then ask it to make forecasts about Europe.

Similarly, you can do 'deep exploration' of claims that human researchers struggle to take seriously. It is a common trope in stories of breakthroughs, particularly in math, that someone got stuck for a long time proving X is true and one day decides on a whim to try to instead prove X is false and does so in hours; this would never happen with LLMs, because you would simply have a search process which tries both equally. This can take an extreme form for really difficult outstanding problems: if a problem like the continuum hypothesis defies all efforts, you could spin up 1000 von Neumann AGIs which have been brainwashed into believing it is false, and then a parallel effort by 1000 brainwashed to believing it is as true as 2+2=4, and let them pursue their research agenda for subjective centuries, and then bring them together to see what important new results they find and how they tear apart the hated enemies' work, for seeding the next iteration.

(These are the sorts of experiments which are why one might wind up running tons of 'ancestor simulations'... There's many more reasons to be simulating past minds than simply very fancy versions of playing The Sims. Perhaps we are now just distant LLM personae being tested about reasoning about the Singularity in one particular scenario involving deep learning counterfactuals, where DL worked, although in the real reality it was Bayesian program synthesis & search.)

Replies from: alexander-gietelink-oldenziel, kromem
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-15T20:12:28.057Z · LW(p) · GW(p)

Beautifully illustrated and amusingly put, sir!

A variant of what you are saying is that AI may once and for all allow us to calculate the true counterfactual     Shapley value of scientific contributions [LW · GW].

( re: ancestor simulations

I think you are onto something here. Compare the Q hypothesis:    

https://twitter.com/dalcy_me/status/1780571900957339771

see also speculations about Zhuangzi hypothesis here  [LW(p) · GW(p)] )

Replies from: gwern
comment by gwern · 2024-05-15T20:35:09.485Z · LW(p) · GW(p)

Yup. Who knows but we are all part of a giant leave-one-out cross-validation computing counterfactual credit assignment on human history? Schmidhuber-em will be crushed by the results.

comment by kromem · 2024-05-15T23:12:19.031Z · LW(p) · GW(p)

While I agree that the potential for AI (we probably need a better term than LLMs or transformers as multimodal models with evolving architectures grow beyond those terms) in exploring less testable topics as more testable is quite high, I'm not sure the air gapping on information can be as clean as you might hope.

Does the AI generating the stories of Napoleon's victory know about the historical reality of Waterloo? Is it using something like SynthID where the other AI might inadvertently pick up on a pattern across the stories of victories distinct from the stories preceding it?

You end up with a turtles all the way down scenario in trying to control for information leakage with the hopes of achieving a threshold that no longer has impact on the result, but given we're probably already seriously underestimating the degree to which correlations are mapped even in today's models I don't have high hopes for tomorrow's.

I think the way in which there's most impact on fields like history is the property by which truth clusters across associated samples whereas fictions have counterfactual clusters. An AI mind that is not inhibited by specialization blindness or the rule of seven plus or minus two and better trained at correcting for analytical biases may be able to see patterns in the data, particularly cross-domain, that have eluded human academics to date (this has been my personal research interest in the area, and it does seem like there's significant room for improvement).

And yes, we certainly could be. If you're a fan of cosmology at all, I've been following Neil Turok's CPT symmetric universe theory closely, which started with the Baryonic asymmetry problem and has tackled a number of the open cosmology questions since. That, paired with a QM interpretation like Everett's ends up starting to look like the symmetric universe is our reference and the MWI branches are variations of its modeling around quantization uncertainties.

(I've found myself thinking often lately about how given our universe at cosmic scales and pre-interaction at micro scales emulates a mathematically real universe, just what kind of simulation and at what scale might be able to be run on a real computing neural network.)

comment by cubefox · 2024-05-15T06:46:36.233Z · LW(p) · GW(p)

This post sounds intriguing, but is largely incomprehensible to me due to not sufficiently explaining the background theories.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-07-16T19:34:54.544Z · LW(p) · GW(p)

What did Yudkoswky get right?

  • The central problem of AI alignment. I am not aware of anything in subsequent work that is not already implicit in Yudkowsky's writing.
  • Short timelines avant le lettre. Yudkowsky was predicting AGI in his lifetime from the very start when most academics, observers, AI scientists, etc considered AGI a fairytale [? · GW].
  • Inherent and irreducible uncertainty of forecasting, foolishness of precise predictions. 
  • The importance of (Pearlian) causality, Solomonoff Induction as theory of formal epistemology, Bayesian statistics, (Shannon) information theory, decision theory [especially UDT-shaped things].  
  • (?nanotech, ?cryonics)
  • if you had a timemachine to go back to 2010 you should buy bitcoin and write Harry Potter fanfiction
Replies from: andrei-alexandru-parfeni, D0TheMath
comment by sunwillrise (andrei-alexandru-parfeni) · 2024-07-16T21:30:19.858Z · LW(p) · GW(p)

From Akash's summary [LW · GW] of the discussion between Conor Leahy and Michael Trazzi on "The Inside View" from ~ 1.5 years ago:

  • A lot of Eliezer’s value as a thinker is that he notices & comprehends antimemes. And he figures out how to communicate them.
  • An antimeme is something that by its very nature resists being known. Most antimemes are just boring—things you forget about. If you tell someone an antimeme, it bounces off them. So they need to be communicated in a special way. Moral intuitions. Truths about yourself. A psychologist doesn’t just tell you “yo, you’re fucked up bro.” That doesn’t work.

In Leahy's own words [LW · GW]:

“Antimemes are completely real. There's nothing supernatural about it. Most antimemes are just things that are boring. So things that are extraordinarily boring are antimemes because they, by their nature, resist you remembering them. And there's also a lot of antimemes in various kinds of sociological and psychological literature. A lot of psychology literature, especially early psychology literature, which is often very wrong to be clear. Psychoanalysis is just wrong about almost everything. But the writing style, the kind of thing these people I think are trying to do is they have some insight, which is an antimeme. And if you just tell someone an antimeme, it'll just bounce off them. That's the nature of an antimeme. So to convey an antimeme to people, you have to be very circuitous, often through fables, through stories you have, through vibes. This is a common thing.

Moral intuitions are often antimemes. Things about various human nature or truth about yourself. Psychologists, don't tell you, "Oh, you're fucked up, bro. Do this." That doesn't work because it's an antimeme. People have protection, they have ego. You have all these mechanisms that will resist you learning certain things. Humans are very good at resisting learning things that make themselves look bad. So things that hurt your own ego are generally antimemes. So I think a lot of what Eliezer does and a lot of his value as a thinker is that he is able, through however the hell his brain works, to notice and comprehend a lot of antimemes that are very hard for other people to understand.

Much of the discussion at the time (example [LW(p) · GW(p)]) focused on the particular application of this idea in the context of the "Death with Dignity" post [LW · GW], but I think this effect was visible much earlier on, most prominently in the Sequences [? · GW] themselves. As I see it, this did not affect the content that was being communicated so much as it did the vibe [LW(p) · GW(p)], the more ineffable, emotional, and hard-to-describe-using-S2 [? · GW] stylistic packaging that enveloped the specific ideas being conveyed. The latter [1], divorced from Eliezer's presentation of them, could be (and often are) thought of as dry or entirely technical, but his writing gave them a certain life that made them rather unforgettable and allowed them to hit much harder (see "How An Algorithm Feels From the Inside" [LW · GW] and "Beyond the Reach of God" [LW · GW] as the standard examples of this).

  1. ^

    Stuff like probability theory [LW · GW], physics (Quantum Mechanics [? · GW] in particular), philosophy of language [? · GW], etc.

comment by Garrett Baker (D0TheMath) · 2024-07-16T19:59:31.732Z · LW(p) · GW(p)

I think I'd agree with everything you say (or at least know what you're looking at as you say it) except for the importance of decision theory. What work are you watching there?

Replies from: habryka4
comment by habryka (habryka4) · 2024-07-16T20:30:32.186Z · LW(p) · GW(p)

As one relevant consideration, I think the topic of "will AI kill all humans" is a question whose answer relies in substantial parts on TDT-ish considerations, and is something that a bunch of value systems I think reasonably care a lot about. Also I think what  superintelligent systems will do will depend a lot on decision-theoretic considerations that seem very hard to answer from a CDT vs. EDT-ish frame.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-07-16T22:33:16.099Z · LW(p) · GW(p)

I think I speak for many when I ask you to please elaborate on this!

Replies from: habryka4
comment by habryka (habryka4) · 2024-07-16T23:36:57.038Z · LW(p) · GW(p)

Oh, I thought this was relatively straightforward and has been discussed a bunch. There are two lines of argument I know for why superintelligent AI, even if unaligned, might not literally kill everyone, but keep some humans alive: 

  1. The AI might care a tiny bit about our values, even if it mostly doesn't share them
  2. The AI might want to coordinate with other AI systems that reached superintelligence to jointly optimize the universe. So in a world where there is only a 1% chance that we align AI systems to our values, then even in unaligned worlds we might end up with AI systems that adopt our values as a 1% mixture in its utility function (and also consequently in those 1% of worlds, we might still want to trade away 99% of the universe to the values that the counterfactual AI systems would have had)

Some places where the second line of argument has been discussed: 

  1. ^

    This is due to:

    • The potential for the AI to be at least a tiny bit "kind" (same as humans probably wouldn't kill all aliens). [1] [LW · GW]
    • Decision theory/trade reasons
  2. ^

    Note that in this comment I’m not touching on acausal trade (with successful humans) or ECL. I think those are very relevant to whether AI systems kill everyone, but are less related to this implicit claim about kindness which comes across in your parables (since acausally trading AIs are basically analogous to the ants who don't kill us because we have power).

Replies from: Raemon
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-03-16T20:08:56.948Z · LW(p) · GW(p)

Feature request: author-driven collaborative editing [CITATION needed] for the Good and Glorious Epistemic Commons.

Often I find myself writing claims which would ideally have citations but I don't know an exact reference, don't remember where I read it, or am simply too lazy to do the literature search. 

This is bad for scholarship is a rationalist virtue. Proper citation is key to preserving and growing the epistemic commons. 

It would be awesome if my lazyness were rewarded by giving me the option to add a [CITATION needed] that others could then suggest (push) a citation, link or short remark which the author (me) could then accept. The contribution of the citator is acknowledged of course. [even better would be if there was some central database that would track citations & links like with crosslinking etc like wikipedia] 

a sort hybrid vigor of Community Notes and Wikipedia if you will. but It's collaborative, not adversarial*

author: blablablabla

sky is blue [citation Needed]

blabblabla

intrepid bibliographer: (push) [1] "I went outside and the sky was blue", Letters to the Empirical Review

 

*community notes on twitter has been a universally lauded concept when it first launched. We are already seeing it being abused unfortunately, often used for unreplyable cheap dunks. I still think it's a good addition to twitter but it does show how difficult it is to create shared agreed-upon epistemics in an adverserial setting. 

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-09-07T12:00:19.877Z · LW(p) · GW(p)

Corrupting influences

The EA AI safety strategy has had a large focus on placing EA-aligned people in A(G)I labs. The thinking was that having enough aligned insiders would make a difference on crucial deployment decisions & longer-term alignment strategy. We could say that the strategy is an attempt to corrupt the goal of pure capability advance & making money towards the goal of alignment. This fits into a larger theme that EA needs to get close to power to have real influence. 

[See also the large donations EA has made to OpenAI & Anthropic. ]

Whether this strategy paid off...  too early to tell.

What has become apparent is that the large AI labs & being close to power have had a strong corrupting influence on EA epistemics and culture. 

  • Many people in EA now think nothing of being paid Bay Area programmer salaries for research or nonprofit jobs.
  •  There has been a huge influx of MBA blabber being thrown around. Bizarrely EA funds are often giving huge grants to for profit organizations for which it is very unclear whether they're really EA-aligned in the long-term or just paying lip service. Highly questionable that EA should be trying to do venture capitalism in the first place. 
  • There is a questionable trend to equate ML skills prestige within capabilities work with the ability to do alignment work.  EDIT: haven't looked at it deeply yet but superfiically impressed by CAIS recent work. seems like an eminently reasonable approach. Hendryx's deep expertise in capabilities work / scientific track record seem to have been key. in general, EA-adjacent AI safety work has suffered from youth, inexpertise & amateurism so makes sense to have more world-class expertise EDITEDIT: i should be careful in promoting work I haven't looked at. I have been told from a source I trust that almost nothing is new in this paper and Hendryx engages in a lot of very questionable self-promotion tactics.
  • For various political reasons there has been an attempt to put x-risk AI safety on a continuum with more mundance AI concerns like it saying bad words. This means there is lots of 'alignment research' that is at best irrelevant, at worst a form of rnsidiuous safetywashing. 

The influx of money and professionalization has not been entirely bad. Early EA suffered much more from virtue signalling spirals, analysis paralysis. Current EA is much more professional, largely for the better. 

Replies from: dmurfet, rhollerith_dot_com, sharmake-farah
comment by Daniel Murfet (dmurfet) · 2023-11-27T18:21:15.336Z · LW(p) · GW(p)

As a supervisor of numerous MSc and PhD students in mathematics, when someone finishes a math degree and considers a job, the tradeoffs are usually between meaning, income, freedom, evil, etc., with some of the obvious choices being high/low along (relatively?) obvious axes. It's extremely striking to see young talented people with math or physics (or CS) backgrounds going into technical AI alignment roles in big labs, apparently maximising along many (or all) of these axes!

Especially in light of recent events I suspect that this phenomenon, which appears too good to be true, actually is.

comment by RHollerith (rhollerith_dot_com) · 2023-09-07T18:24:56.800Z · LW(p) · GW(p)

There is a questionable trend to equate ML skills with the ability to do alignment work.

Yes!

Replies from: thomas-kwa
comment by Thomas Kwa (thomas-kwa) · 2023-09-07T18:58:44.097Z · LW(p) · GW(p)

I'm not too concerned about this. ML skills are not sufficient to do good alignment work, but they seem to be very important for like 80% of alignment work and make a big difference in the impact of research (although I'd guess still smaller than whether the application to alignment is good)

  • Primary criticisms of Redwood [LW · GW] involve their lack of experience in ML
  • The explosion of research in the last ~year is partially due to an increase in the number of people in the community who work with ML. Maybe you would argue that lots of current research is useless, but it seems a lot better than only having MIRI around
  • The field of machine learning at large is in many cases solving easier versions of problems we have in alignment, and therefore it makes a ton of sense to have ML research experience in those areas. E.g. safe RL is how to get safe policies when you can optimize over policies and know which states/actions are safe; alignment can be stated as a harder version of this where we also need to deal with value specification, self-modification, instrumental convergence etc.
Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-09-07T20:45:38.780Z · LW(p) · GW(p)

I mostly agree with this.

I should have said 'prestige within capabilities research' rather than ML skills which seems straightforwardly useful. The former is seems highly corruptive.

comment by Noosphere89 (sharmake-farah) · 2023-09-09T14:53:37.231Z · LW(p) · GW(p)

There is a questionable trend to equate ML skills with the ability to do alignment work.

I'd arguably say this is good, primarily because I think EA was already in danger of it's AI safety wing becoming unmoored from reality by ignoring key constraints, similar to how early Lesswrong before the deep learning era around 2012-2018 turned out to be mostly useless due to how much everything was stated in a mathematical way, and not realizing how many constraints and conjectured constraints applied to stuff like formal provability, for example..

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-04T13:34:31.176Z · LW(p) · GW(p)

Entropy and AI Forecasting 

Until relatively recently (2018-2019?) I did not seriously entertain the possibility that AGI in our lifetime was possible. This was a mistake, an epistemic error. A rational observer calmly and objectively considering the evidence for AI progress over the prior decades - especially in the light of rapid progress in deep learning - should have come to the reasonable position that AGI within 50 years was a serious possibility (>10%). 

AGI plausibly arriving in our lifetime was a reasonable position. Yet this possibility was almost universally ridiculed or ignored or by academics and domain experts. One can find quite funny interview with AI experts on Lesswrong from 15 years ago. The only AI expert agreeing with the Yudkowskian view of AI in our lifetime was Jurgen Schmidthuber. The other dozen AI experts denied it as unknowable or even denied the hypothetical possibility of AGI. 

Yudkowsky earns a ton of Bayes points for anticipating the likely arrival of AGI in our lifetime long before the deep learning took off. 

**************************

We are currently experiencing a rapid AI takeoff, plausibly culminating in superintelligence by the end of this decade. I know of only two people who anticipated something like what we are seeing far ahead of time; Hans Moravec and Jan Leike  Shane Legg*. Both forecast fairly precise dates decades before it happened - and the reasons why they thought it would happen are basically the reasons it did (i.e. Moravec very early on realized the primacy of compute). Moreover, they didn't forecast a whole lot of things that didn't happen (like Kurzweil).

Did I make an epistemic error by not believing them earlier? Well for starters I wasn't really plugged in to the AI scene so I hadn't heard of them or their views. But suppose I did; should I have beieved them? I'd argue I shouldn't give their view back then more a little bit of credence. 

Entropy is a mysterious physics word for irreducible uncertainty; the uncertainty that remains about the future even after accounting for all the data. In hindsight, we can say that massive GPU training on next-token prediction of all internet text data was (almost**) all you need for AGI. But was this forecasteable?

For every Moravec and Leike Legg who turns out to be extraordinairly right in forecasting the future there re dozens that weren't. Even in 2018 when the first evidence for strong scaling laws on text-data was being published by Baidu I'd argue that an impartial observer should have only updated a moderate amount. Actually even OpenAI itself wasn't sold on unsupervised learning on textdata until early gpt showed signs of life - they thought (like many other players in the field, e.g. DeepMind) that RL (in diverse environments) was the way to go. 


To me the takeaway is that explicit forecasting can be useful but it is exactly the blackswan events that are irreducibly uncertain (high entropy) that move history. 


 

*the story is that Leike Legg's timelines have been 2030 for the past two decades. 

** regular readers will know my beef with the pure scaling hypothesis. 

Replies from: interstice, bogdan-ionut-cirstea, sharmake-farah
comment by interstice · 2024-10-04T18:17:23.963Z · LW(p) · GW(p)

I know of only two people who anticipated something like what we are seeing far ahead of time; Hans Moravec and Jan Leike

I didn't know about Jan's AI timelines. Shane Legg also had some decently early predictions of AI around 2030(~2007 was the earliest I knew about)

Replies from: mark-xu, alexander-gietelink-oldenziel
comment by Mark Xu (mark-xu) · 2024-10-04T21:03:43.762Z · LW(p) · GW(p)

shane legg had 2028 median back in 2008, see e.g. https://e-discoveryteam.com/2023/11/17/shane-leggs-vision-agi-is-likely-by-2028-as-soon-as-we-overcome-ais-senior-moments/

Replies from: interstice
comment by interstice · 2024-10-05T04:25:06.885Z · LW(p) · GW(p)

That's probably the one I was thinking of.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-05T08:13:57.240Z · LW(p) · GW(p)

Oh no uh-oh I think I might have confused Shane Legg with Jan Leike

comment by Bogdan Ionut Cirstea (bogdan-ionut-cirstea) · 2024-10-05T10:05:32.622Z · LW(p) · GW(p)

Fwiw, in 2016 I would have put something like 20% probability on what became known as 'the scaling hypothesis'. I still had past-2035 median timelines, though. 

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-05T10:58:46.236Z · LW(p) · GW(p)

What did you mean exactly in 2016 by the scaling hypothesis ?

Having past 2035 timelines and believing in the pure scaling maximalist hypothesis (which fwiw i don't believe in for reasons i have explained elsewhere) are in direct conflict so id be curious if you could more exactly detail your beliefs back then.

Replies from: bogdan-ionut-cirstea
comment by Bogdan Ionut Cirstea (bogdan-ionut-cirstea) · 2024-10-05T11:21:03.992Z · LW(p) · GW(p)

What did you mean exactly in 2016 by the scaling hypothesis ?

Something like 'we could have AGI just by scaling up deep learning / deep RL, without any need for major algorithmic breakthroughs'.

Having past 2035 timelines and believing in the pure scaling maximalist hypothesis (which fwiw i don't believe in for reasons i have explained elsewhere) are in direct conflict so id be curious if you could more exactly detail your beliefs back then.

I'm not sure this is strictly true, though I agree with the 'vibe'. I think there were probably a couple of things in play:

  • I still only had something like 20% on scaling, and I expected much more compute would likely be needed, especially in that scenario, but also more broadly (e.g. maybe something like the median in 'bioanchors' - 35 OOMs of pretraining-equivalent compute, if I don't misremember; though I definitely hadn't thought very explicitly about how many OOMs of compute at that time) - so I thought it would probably take decades to get to the required amount of compute.
  • I very likely hadn't thought hard and long enough to necessarily integrate/make coherent my various beliefs. 
  • Probably at least partly because there seemed to be a lot of social pressure from academic peers against even something like '20% on scaling', and even against taking AGI and AGI safety seriously at all. This likely made it harder to 'viscerally feel' what some of my beliefs might imply, and especially that it might happen very soon (which also had consequences in delaying when I'd go full-time into working on AI safety; along with thinking I'd have more time to prepare for it, before going all in).
comment by Noosphere89 (sharmake-farah) · 2024-10-04T16:33:25.398Z · LW(p) · GW(p)

Yeah, I do think that Moravec and Leike got the AI situation most correct, and yeah people were wrong to dismiss Yudkowsky for having short timelines.

This was the thing they got most correct, which is interesting because unfortunately, Yudkowsky got almost everything else incorrect about how superhuman AIs would work, and also got the alignment situation very wrong as well, which is very important to take note of.

LW in general got short timelines and the idea that AI will probably be the biggest deal in history correct, but went wrong in assuming they knew well about how AI would eventually work (remember the times when Eliezer Yudkowsky dismissed neural networks working for capabilities instead of legible logic?) and also got the alignment situation very wrong, due to way overcomplexifying human values and relying on the evopsych frame way too much for human values, combined with not noticing that the differences between humans and evolution that mattered for capabilities also mattered for alignment.

I believe a lot of the issue comes down to incorrectly conflating the logical possibility of misalignment with the probability of misalignement being high enough that we should take serious action, and the interlocutors they talked with often denied the possibility that misalignment could happen at all, but LWers then didn't realize that reality doesn't grade on a curve, and though their arguments were better than their interlocutors, that didn't mean they were right.

Replies from: alexander-gietelink-oldenziel, quetzal_rainbow
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-04T17:00:20.149Z · LW(p) · GW(p)

Yudkowsky didnt dismiss neural networks iirc. He just said that there were a lot of different approaches to AI and from the Outside View it didnt seem clear which was promising - and plausibly on an Inside View it wasnt very clear that aritificial neural networks were going to work and work so well.

Re:alignment I dont follow. We dont know who will be proved ultimately right on alignment so im not sure how you can make such strong statements about whether Yudkowsky was right or wrong on this aspect.

We havent really gained that much bits on this question and plausibly will not gain many until later (by which time it might be too late if Yudkowsky is right).

I do agree that Yudkowsky's statements occasionally feel too confidently and dogmatically pessimistic on the question of Doom. But I would argue that the problem is that we simply dont know well because of irreducible uncertainty - not that Doom is unlikely.

Replies from: sharmake-farah
comment by Noosphere89 (sharmake-farah) · 2024-10-04T17:43:45.931Z · LW(p) · GW(p)

Mostly, I'm annoyed by how much his argumentation around alignment matches the pattern of dismissing various approaches to alignment using similar reasoning to how he dismissed neural networks:

Even if it was correct to dismiss neural networks years ago, it isn't now, so it's not a good sign that the arguments rely on this issue:

https://www.lesswrong.com/posts/wAczufCpMdaamF9fy/my-objections-to-we-re-all-gonna-die-with-eliezer-yudkowsky#HpPcxG9bPDFTB4i6a [LW(p) · GW(p)]

I am going to argue that we do have quite a lot of bits on alignment, and the basic argument can be summarized like this:

Human values are much less complicated than people thought, and also more influenced by data than people thought 15-20 years ago, and thus much, much easier to specify than people thought 15-20 years ago.

That's the takeaway I have from current LLMs handling human values, and I basically agree with Linch's summary of Matthew Barnett's post on the historical value misspecification argument of what that means in practice for alignment:

https://www.lesswrong.com/posts/i5kijcjFJD6bn7dwq/evaluating-the-historical-value-misspecification-argument#N9ManBfJ7ahhnqmu7 [LW(p) · GW(p)]

It's not about LLM safety properties, but about what has been revealed about our values.

Another way to say it is that we don't need to reverse-engineer social instincts for alignment, contra @Steven Byrnes [LW · GW], because we can massively simplify what the social instinct parts of our brain that contribute to alignment are doing in code, because while the mechanisms for how humans get their morality and not be psychopaths are complicated, it doesn't matter, because we can replicate it's function with much simpler code and data, and go to a more blank-slate design for AIs:

https://www.lesswrong.com/posts/PTkd8nazvH9HQpwP8/building-brain-inspired-agi-is-infinitely-easier-than#If_some_circuit_in_the_brain_is_doing_something_useful__then_it_s_humanly_feasible_to_understand_what_that_thing_is_and_why_it_s_useful__and_to_write_our_own_CPU_code_that_does_the_same_useful_thing_ [LW · GW]

(A similar trick is one path to solving robotics for AIs, but note this is only one part, it might be that the solution routes through a different mechanism).

Really, I'm not mad about his original ideas, because they might have been correct, and it wasn't obviously incorrect, I'm just mad that he didn't realize that he had to update to reality more radically than he had realized, and seems to conflate the bad argument for AI will understand our values, therefore it's safe, with the better argument that LLMs show it's easier to specify values without drastically wrong results, and that it's not a complete solution to alignment, but a big advance on outer alignment in the usual dichotomy.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-04T17:59:26.470Z · LW(p) · GW(p)

It's a plausible argument imho. Time will tell.

To my mind an important dimension, perhaps the most important dimensions is how values be evolve under reflection.

It's quite plausible to me that starting with an AI that has pretty aligned values it will self-reflect into evil. This is certainly not unheard of in the real world (let alone fiction!). Of course it's a question about the basin of attraction around helpfulness and harmlessness. I guess I have only weak priors on what this might look like under reflection, although plausibly friendliness is magic.

Replies from: D0TheMath, sharmake-farah
comment by Garrett Baker (D0TheMath) · 2024-10-04T18:14:49.540Z · LW(p) · GW(p)

It's quite plausible to me that starting with an AI that has pretty aligned values it will self-reflect into evil.

I disagree, but could be a difference in definition of what "perfectly aligned values" means. Eg if the AI is dumb (for an AGI) and in a rush, sure. If its a superintelligence already, even in a rush, seems unlikely. [edit:] If we have found an SAE feature which seems to light up for good stuff, and down for bad stuff 100% of the time, then we clamp it, then yeah, that could go away on reflection.

comment by Noosphere89 (sharmake-farah) · 2024-10-04T18:13:31.089Z · LW(p) · GW(p)

Another way to say it is how values evolve in OOD situations.

My general prior, albeit reasonably weak is that the best single way to predict how values evolve is looking at their data sources, as well as what data they received up to now, and the second best way to predict it is looking at what their algorithms are, especially for social situations, and that most of the other factors don't matter nearly as much.

comment by quetzal_rainbow · 2024-10-04T18:18:37.429Z · LW(p) · GW(p)

Yudkowsky got almost everything else incorrect about how superhuman AIs would work,

I think this statement is incredibly overconfident, because literally nobody knows how superhuman AI would work.

And, I think, this is general shape of problem: incredible number of people got incredibly overindexed on how LLMs worked in 2022-2023 and drew conclusions which seem to be plausible, but not as probable as these people think.

Replies from: sharmake-farah
comment by Noosphere89 (sharmake-farah) · 2024-10-04T18:30:11.562Z · LW(p) · GW(p)

Okay, I talked more on what conclusions we can draw from LLMs that actually generalize to superhuman AI here, so go check that out:

https://www.lesswrong.com/posts/tDkYdyJSqe3DddtK4/alexander-gietelink-oldenziel-s-shortform#mPaBbsfpwgdvoK2Z2 [LW(p) · GW(p)]

The really short summary is human values are less complicated and more dependent on data than people thought, and we can specify our values rather easily without it going drastically wrong:

This is not a property of LLMs, but of us.

Replies from: D0TheMath
comment by Garrett Baker (D0TheMath) · 2024-10-04T21:51:39.362Z · LW(p) · GW(p)

here

is that supposed to be a link?

Replies from: sharmake-farah, sharmake-farah
comment by Noosphere89 (sharmake-farah) · 2024-10-04T23:02:06.206Z · LW(p) · GW(p)

I rewrote the comment to put the link immediately below the first sentence.

comment by Noosphere89 (sharmake-farah) · 2024-10-04T21:53:49.347Z · LW(p) · GW(p)

The link is at the very bottom of the comment.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-07-24T16:06:24.202Z · LW(p) · GW(p)

Crypticity, Reverse Epsilon Machines and the Arrow of Time?

[see https://arxiv.org/abs/0902.1209 ]

Our subjective experience of the arrow of time is occasionally suggested to be an essentially entropic phenomenon. 

This sounds cool and deep but crashes headlong into the issue that the entropy rate and the excess entropy of any stochastic process is time-symmetric. I find it amusing that despite hearing this idea often from physicists and the like apparently this rather elementary fact has not prevented their storycrafting. 

Luckily, computational mechanics provides us with a measure that is not time symmetric: the stochastic complexity of the epsilon machine 

For any stochastic process we may also consider the epsilon machine of the reverse process, in other words the machine that predicts the past based on the future. This can be a completely different machine whose reverse stochastic complexity  is not equal to 

Some processes are easier to predict forward than backward. For example, there is considerable evidence that language is such a process. If the stochastic complexity and the reverse stochastic complexity differ we speak of a causally assymetric process. 

Alec Boyd pointed out to me that the classic example of a glass falling of a table is naturally thought of in these terms. The forward process is easy to describe while the backward process is hard to describe where easy and hard are meant in the sense of stochastic complexity: bits needed to specify the states of perfect minimal predictor, respectively retrodictor. 

rk. note that time assymmetry is a fundamentally stochastic phenomenon. THe underlyiing (let's say classicially deterministic) laws are still time symmetric. 

The hypothesis is then: many, most macroscopic processes of interest to humans, including other agents are fundamentally such causally assymetric (and cryptic) processes. 

Replies from: Lblack
comment by Lucius Bushnaq (Lblack) · 2024-07-24T17:20:14.085Z · LW(p) · GW(p)

This sounds cool and deep but crashes headlong into the issue that the entropy rate and the excess entropy of any stochastic process is time-symmetric.
 

It's time symmetric around a starting point  of low entropy. The further  is from , the more entropy you'll have, in either direction. The absolute value  is what matters.


In this case,  is usually taken to be the big bang.  So the further in time you are from the big bang, the less the universe is like a dense uniform soup with little structure that needs description, and the higher your entropy will be. That's how you get the subjective perception of temporal causality. 

Presumably, this would hold to the other side of  as well, if there is one. But we can't extrapolate past , because close to  everything gets really really energy dense, so we'd need to know how to do quantum gravity to calculate what the state on the other side might look like.  So we can't check that.  And the notion of time as we're discussing it here might break down at those energies anyway.

Replies from: cubefox
comment by cubefox · 2024-07-26T02:20:28.249Z · LW(p) · GW(p)

See also the Past Hypothesis. If we instead take a non-speculative starting point as , namely now, we could no longer trust our memories, including any evidence we believe to have about the entropy of the past being low, or about physical laws stating that entropy increases with distance from . David Albert therefore says doubting the Past Hypothesis would be "epistemically unstable".

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-11-27T12:48:27.739Z · LW(p) · GW(p)

Pockets of Deep Expertise 

Why am I so bullish on academic outreach? Why do I keep hammering on 'getting the adults in the room'? 

It's not that I think academics are all Super Smart. 

I think rationalists/alignment people correctly ascertain that most professors don't have much useful to say about alignment & deep learning and often say silly things. They correctly see that much of AI congress is fueled by labs and scale not ML academia. I am bullish on non-ML academia, especially mathematics, physics and to a lesser extent theoretical CS, neuroscience, some parts of ML/ AI academia. This is because while I think 95 % of academia is bad and/or useless there are Pockets of Deep Expertise. Most questions in alignment are close to existing work in academia in some sense - but we have to make the connection!

A good example is 'sparse coding' and 'compressed sensing'. Lots of mech.interp has been rediscovering some of the basic ideas of sparse coding. But there is vast expertise in academia about these topics. We should leverage these!

Other examples are singular learning theory, computational mechanics, etc

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-01T11:07:13.613Z · LW(p) · GW(p)

I have an embarrasing confession to make. I don't understand why PvsNP is so hard. 

[I'm in good company since apparently Richard Feynmann couldn't be convinced it was a serious open problem.] 

I think I understand PvsNP and its many variants like existence of one-way function is about computational hardness of certain tasks. It is surprising that we have such strong intuitions that some tasks are computationally hard but we fail to be able to prove it!

Of course I don't think I can prove it and I am not foolish enough to spend significant amount of time on trying to prove it. I still would like to understand the deep reasons  why it's so hard to prove computational hardness results. That means I'd like to understand why certain general proof strategies are impossible or very hard. 

There is an old argument by Shannon that proves that almost every* Boolean function has exponential circuit depth. This is a simple counting argument. Basically, there are exponentially many more Boolean functions than there are circuits. It's hard to give explicit examples of  computationally hard functions** but we can easily show they are plentiful. 

This would seem to settle the matter of existence of computationally hard functions. I believe the rattlesnake in the grass is that the argument only proves that Boolean functions are computationally hard to compute in terms of Boolean circuits but general computable algorithms are more expressive? I am not entirely sure about this. 

 I have two confusions about this: 

Confusion #1: General algorithms would need to make use of some structure. They aren't magic. Can't we solve that if you could do this in general you would need to effectively 'simulate' these Boolean circuits which would reduce the proof to Shannon-like counting argument?

Confusion #2: Why couldn't we make similar counting arguments for Turing machines?

Shannon's argument is very similar to the basic counting argument in algorithmic information theory, showing that most strings are K-incompressible. 

Rk. There are the famous 'proof barriers' to a proof of PvsNP like natural proofs and algebraization. I don't understand these ideas - perhaps they can shed some light on the matter. 

 

@Dalcy [LW · GW

*the complement is exponentially sparse

** parity functions? 

Replies from: quetzal_rainbow, kh, dmitry-vaintrob, tailcalled, sharmake-farah, Mo Nastri, TsviBT, dmitry-vaintrob
comment by quetzal_rainbow · 2024-10-01T12:11:44.489Z · LW(p) · GW(p)

I'm just computational complexity theory enthusiast, but my opinion is that P vs NP centered explanation of computational complexity is confusing. Explanation of NP should happen in the very end of the course.

There is nothing difficult in proving that computationally hard functions exist: time hierarchy theorem implies that, say, P is not equal EXPTIME. Therefore, EXPTIME is "computationally hard". What is difficult is to prove that very specific class of problems which have zero-error polynomial-time verification algorithms is "computationally hard".

comment by Kaarel (kh) · 2024-10-02T06:21:40.188Z · LW(p) · GW(p)

Confusion #2: Why couldn't we make similar counting arguments for Turing machines?

I guess a central issue with separating NP from P with a counting argument is that (roughly speaking) there are equally many problems in NP and P. Each problem in NP has a polynomial-time verifier, so we can index the problems in NP by polytime algorithms, just like the problems in P.

in a bit more detail: We could try to use a counting argument to show that there is some problem with a (say) time verifier which does not have any (say) time solver. To do this, we'd like to say that there are more verifier problems than algorithms. While I don't really know how we ought to count these (naively, there are of each), even if we had some decent notion of counting, there would almost certainly just be more algorithms than verifiers (since the verifiers are themselves algorithms).

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-02T10:42:26.413Z · LW(p) · GW(p)

Thank you Kaarel - this the kind of answer I was after.

comment by Dmitry Vaintrob (dmitry-vaintrob) · 2024-10-01T17:49:18.195Z · LW(p) · GW(p)

Looking at this again, I'm not sure I understand the two confusions. P vs. NP isn't about functions that are hard to compute (they're all polynomially computable), rather functions that are hard to invert, or pairs of easily computable functions that hard to prove are equal/not equal to each other. The main difference between circuits and Turing machines is that circuits are finite and bounded to compute whereas the halting time of general Turing machines is famously impossible to determine. There's nothing special about Boolean circuits: they're an essentially complete model of what can be computed in polynomial time (modulo technicalities)

Replies from: dmitry-vaintrob
comment by Dmitry Vaintrob (dmitry-vaintrob) · 2024-10-01T18:03:22.208Z · LW(p) · GW(p)

In particular, it's not hard to produce a computable function that isn't given by a polynomial-sized circuit (parity doesn't work as it's polynomial, but you can write one down using diagonalization -- it would be very long to compute, but computable in some suitably exponentially bounded time). But P vs. NP is not about this: it's a statement that exists fully in the world of polynomially computable functions.

comment by tailcalled · 2024-10-01T13:08:18.271Z · LW(p) · GW(p)

We know there are difficult computational problems. P vs NP is more narrow than that; it's sometimes phrased as "are there problems that are not difficult to verify but difficult to solve?", where "difficult" means that it cannot be done in asymptotically polynomial time.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-01T16:22:31.279Z · LW(p) · GW(p)

Yes, I am familiar with the definition of PvsNP. That's not what I am asking.

Replies from: sharmake-farah
comment by Noosphere89 (sharmake-farah) · 2024-10-01T16:47:17.588Z · LW(p) · GW(p)

The point is that you can't use the result that there exists a hard function, since all you know is that the function is hard, not whether it's in NP, which is a basic problem for your scheme.

Your counting argument for Turing Machines also likely have this problem, and even if not, I see no reason why I couldn't relativize the results, which is a no-go for P vs NP proof attempts.

comment by Noosphere89 (sharmake-farah) · 2024-10-01T16:43:48.635Z · LW(p) · GW(p)

Basically, there are 3 main barriers to proving P not equaling NP.

One, you have to actually show that there exists a hard function that isn't in P, and it's not enough to prove that there are exponentially many hard functions, because it might be that a circuit computing an NP-complete problem has a linear time bound.

And natural proofs argue that unless cryptographically hard functions don't exist, the common way to prove circuit lower bounds also can't prove P vs NP (Technical details are below:)

https://en.wikipedia.org/wiki/Natural_proof

Also, both of the strategies cannot relativize or algebrize, where relativization means that if we give a fixed oracle tape O consisting of a single problem you can solve instantly to all parties doesn't change the conclusion for all oracle tapes O.

Many results like this, including possibly your attempts to prove via counting arguments almost certainly relativize, and even if they don't, they algebrize, and the technical details are below for algebrization are here, since I already explained relativization above.

https://www.scottaaronson.com/papers/alg.pdf

But that's why proving P vs NP is so hard technically.

comment by Mo Putera (Mo Nastri) · 2024-10-01T15:45:14.295Z · LW(p) · GW(p)

You might be interested in Scott Aaronson's thoughts on this in section 4: Why Is Proving P != NP Difficult?, which is only 2 pages. 

Replies from: dmitry-vaintrob
comment by Dmitry Vaintrob (dmitry-vaintrob) · 2024-10-01T16:05:02.454Z · LW(p) · GW(p)

looks like you referenced the same paper before me while I was making my comment :)

Replies from: Mo Nastri
comment by Mo Putera (Mo Nastri) · 2024-10-02T07:10:51.222Z · LW(p) · GW(p)

Ha, that's awesome. Thanks for including the screenshot in yours :) Scott's "invisible fence" argument was the main one I thought of actually.

comment by Dmitry Vaintrob (dmitry-vaintrob) · 2024-10-01T15:59:14.367Z · LW(p) · GW(p)

Yeah I think this is a good place to probe assumptions, and it's probably useful to form world models where you probability of P = NP is nonzero (I also like doing this for inconsistency of logic). I don't have an inside view, but like Scott Aaronson on this: https://www.scottaaronson.com/papers/pnp.pdf:
 

Replies from: sharmake-farah
comment by Noosphere89 (sharmake-farah) · 2024-10-01T17:10:35.776Z · LW(p) · GW(p)

My real view on P vs NP is that at this point, I think P almost certainly not equal to NP, and that any solving of NP-complete problems efficiently to the standard of complexity theorists requires drastically changing the model of computation, which corresponds to drastic changes in our physics assumptions like time travel actually working according to Deutsch's view (and there being no spurious fixed-points).

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-20T12:15:55.481Z · LW(p) · GW(p)

Why no prediction markets for large infrastructure projects?

Been reading this excellent piece on why prediction markets aren't popular. They say that without subsidies prediction markets won't be large enough; the information value of prediction markets is often nog high enough. 

Large infrastructure projects undertaken by governments, and other large actors often go overbudget, often hilariously so: 3x,5x,10x or more is not uncommon, indeed often even the standard.

One of the reasons is that government officials deciding on billion dollar infrastructure projects don't have enough skin in the game. Politicians are often not long enough in office to care on the time horizons of large infrastructure projects. Contractors don't gain by being efficient or delivering on time. To the contrary, infrastructure projects are huge cashcows. Another problem is that there are often far too many veto-stakeholders. All too often the initial bid is wildly overoptimistic. 

Similar considerations apply to other government projects like defense procurement or IT projects.

Okay - how to remedy this situation? Internal prediction markets theoretically could prove beneficial. All stakeholders & decisionmakers are endowed with vested equity with which they are forced to bet on building timelines and other key performance indicators. External traders may also enter the market, selling and buying the contracts. The effective subsidy could be quite large. Key decisions could save billions. 

In this world, government officials could gain a large windfall which may be difficult to explain to voters. This is a legitimate objection. 

A very simple mechanism would simply ask people to make an estimate on the cost C and the timeline T for completion.  Your eventual payout would be proportional to how close you ended up to the real C,T compared to the other bettors. [something something log scoring rule is proper]. 

Replies from: jeremy-gillen
comment by Jeremy Gillen (jeremy-gillen) · 2024-05-20T16:46:36.086Z · LW(p) · GW(p)

Doesn't the futarchy hack come up here? Contractors will be betting that competitors timelines and cost will be high, in order to get the contract. 

Replies from: carl-feynman
comment by Carl Feynman (carl-feynman) · 2024-05-20T17:13:51.172Z · LW(p) · GW(p)

The standard reply is that investors who know or suspect that the market is being systematically distorted will enter the market on the other side, expecting to profit from the distortion. Empirically, attempts to deliberately sway markets in desired directions don’t last very long.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-09-28T18:50:18.956Z · LW(p) · GW(p)

Fractal Fuzz: making up for size

GPT-3 recognizes 50k possible tokens. For a 1000 token context window that means there are  possible prompts. Astronomically large. If we assume the output of a single run of gpt is 200 tokens then for each possible prompt there are  possible continuations. 

GPT-3 is probabilistic, defining for each possible prompt  () a distribution  on a set of size , in other words a  dimensional space. [1]

Mind-boggingly large. Compared to these numbers the amount of data (40 trillion tokens??) and the size of the model (175 billion parameters) seems absolutely puny in comparison.

I won't be talking about the data, or 'overparameterizations' in this short, that is well-explained by Singular Learning Theory. Instead, I will be talking about nonrealizability.

Nonrealizability & the structure of natural data

Recall the setup of (parametric) Bayesian learning: there is a sample space , a true distribution  on  and a parameterized family of probability distributions .

It is often assumed that the true distribution is 'realizable', i.e.  for some . Seeing the numbers in the previous section this assumption seems dubious but the situation becomes significantly easier to analyze, both conceptually and mathematically when we assume realizability.

Conceptually, if the space of possible true distributions is very large compared to the space of model parameters we may ask: how do we know that the true distribution is in the model (or can be well-approximated by it?). 

One answer one hears often is the 'universal approximation theorem' (i.e. the Stone-Weierstrass theorem). I'll come back to this shortly.

Another point of view is that real data sets are actually localized in a very low dimensional subset of all possible data .[2] Following this road leads to theories of lossless compressions, cf sparse coding and compressed sensing which are of obvious important to interpreting modern neural networks. 

That is lossless compression, but another side of the coin is lossy compression.

Fractals and lossy compression

GPT-3 has 175 billion parameters, but the space of possible continuations is many times larger  . Even if we sparse coding implies that the effective dimensionality is much smaller - is it really small enough?

Whenever we have a lower-dimensional subspace  of a larger dimensional subspace there are points  in the larger dimensional space that are very (even arbitrarily) far from . This is easy to see in the linear case but also true if  is more like a manifold[3]- the volume of a lower dimensional space is vanishly small compared to the higher dimensional space. It's a simple mathematical fact that can't be denied!

Unless...  is a fractal. 

This is from Marzen & Crutchfield's "Nearly Maximally Predictive Features and Their Dimensions". The setup is Crutchfield Computational Mechanics, whose central characters are Hidden Markov Models. I won't go into the details here [but give it a read!].

The conjecture is the following: a 'good architecture' defines a model space  that is effectively a fractal in much larger-dimensional space of  realistic data distributions such that for any possible true distribution  the KL-divergence  for some small  . 

Grokking

Phase transitions in loss when varying model size are designated 'grokking'. We can combine the fractal data manifold hypothesis with a SLT-perspective: as we scale up the model, the set of optimal parameter  become better and better. It could happen that the model size gets big enough that it includes a whole new phase, meaning a  with radically lower loss  and higher 

EDIT: seems I'm confused about the nomenclature. Grokking doesn't refer to phase transitions in model size, but in training and data size. 

EDIT2: Seems I'm not crazy. Thanks to Matt Farugia to pointing me towards this result: neural networks are strongly nonconvex (i.e. fractal)

 

Image

EDIT: seems to me that there is another point of contention in which universal approximation theorems (Stone-Weierstrass theorem) are misleading. The stone-Weierstrass applies to a subalgebra of the continuous functions. Seems to me that in the natural parameterization ReLU neural networks aren't a subalgebra of the continuous functions (see also the nonconvexity above). 
 

To think about: information dimension 

  1. ^

    Why the -1? Think visually about the set of distributions on 3 points. Hint: it's a solid triangle ('2- simplex') 

  2. ^

    I think MLers call this the 'data manifold'?

  3. ^

    In the mathematical sense of smooth manifold, not the ill-defined ML notion of 'data manifold'.

Replies from: zfurman
comment by Zach Furman (zfurman) · 2023-09-28T20:11:00.575Z · LW(p) · GW(p)

Very interesting, glad to see this written up! Not sure I totally agree that it's necessary for  to be a fractal? But I do think you're onto something.

In particular you say that "there are points  in the larger dimensional space that are very (even arbitrarily) far from ," but in the case of GPT-4 the input space is discrete, and even in the case of e.g. vision models the input space is compact. So the distance must be bounded.

Plus if you e.g. sample a random image, you'll find there's usually a finite distance you need to travel in the input space (in L1, L2, etc) until you get something that's human interpretable (i.e. lies on the data manifold). So that would point against the data manifold being dense in the input space.

But there is something here, I think. The distance usually isn't that large until you reach a human interpretable image, and it's quite easy to perturb images slightly to have completely different interpretations (both to humans and ML systems). A fairly smooth data manifold wouldn't do this. So my guess is that the data "manifold" is in fact not a manifold globally, but instead has many self-intersections and is singular. That would let it be close to large portions of input space without being literally dense in it. This also makes sense from an SLT perspective. And IIRC there's some empirical evidence that the dimension of the data "manifold" is not globally constant.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-09-28T20:30:27.760Z · LW(p) · GW(p)

The input and output spaces etc  are all discrete but the spaces of distributions  on those spaces are infinite (but still finite-dimensional). 

It depends on what kind of metric one uses, compactness assumptions etc whether or not you can be arbitrarily far. I am being rather vague here. For instance, if you use the KL-divergence, then  is always bounded -  indeed it equals the entropy of the true distribution !

I don't really know what ML people mean by the data manifold so won't say more about that. 

I am talking about the space  of parameter values of a conditional probability distribution 

 I think that  having nonconstant local dimension doesn't seem that relevant since the largest dimensional subspace would dominate?

Self-intersections and singularities could certainly occur here. (i) singularities in the SLT sense have to do with singularities in the level sets of the KL-divergence (or loss function)  - don't see immediately how these are related to the singularities that you are talking about here (ii) it wouldn't increase the dimensionality (rather the opposite). 

The fractal dimension is important basically because of space-filling curves : a space that has a low-dimensional parameterization can nevertheless have a very large effective dimensions when embedded fractally into a larger-dimensional space. These embeddings can make a low-dimensional parameterization effectively have higher dimension. 

Replies from: zfurman
comment by Zach Furman (zfurman) · 2023-09-28T21:42:17.828Z · LW(p) · GW(p)

Sorry, I realized that you're mostly talking about the space of true distributions and I was mainly talking about the "data manifold" (related to the structure of the map  for fixed ). You can disregard most of that.

Though, even in the case where we're talking about the space of true distributions, I'm still not convinced that the image of  under  needs to be fractal. Like, a space-filling assumption sounds to me like basically a universal approximation argument - you're assuming that the image of  densely (or almost densely) fills the space of all probability distributions of a given dimension. But of course we know that universal approximation is problematic and can't explain what neural nets are actually doing for realistic data.

Replies from: alexander-gietelink-oldenziel, alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-09-28T22:03:31.884Z · LW(p) · GW(p)

Obviously this is all speculation but maybe I'm saying that the universal approximation theorem implies that neural architectures are fractal in space of all distributtions (or some restricted subset thereof)?

Curious what's your beef with universal approximation? Stone-weierstrass isn't quantitative - is that the reason?

If true it suggest the fractal dimension (probably related to the information dimension I linked to above) may be important.

Replies from: zfurman
comment by Zach Furman (zfurman) · 2023-09-28T23:15:44.992Z · LW(p) · GW(p)

Obviously this is all speculation but maybe I'm saying that the universal approximation theorem implies that neural architectures are fractal in space of all distributtions (or some restricted subset thereof)?


Oh I actually don't think this is speculation, if (big if) you satisfy the conditions for universal approximation then this is just true (specifically that the image of  is dense in function space). Like, for example, you can state Stone-Weierstrass as: for a Hausdorff space X, and the continuous functions under the sup norm , the Banach subalgebra of polynomials is dense in . In practice you'd only have a finite-dimensional subset of the polynomials, so this obviously can't hold exactly, but as you increase the size of the polynomials, they'll be more space-filling and the error bound will decrease.

Curious what's your beef with universal approximation? Stone-weierstrass isn't quantitative - is that the reason?

The problem is that the dimension of  required to achieve a given  error bound grows exponentially with the dimension  of your underlying space . For instance, if you assume that weights depend continuously on the target function, -approximating all  functions on  with Sobolev norm  provably takes at least  parameters (DeVore et al.). This is a lower bound.

So for any realistic  universal approximation is basically useless - the number of parameters required is enormous. Which makes sense because approximation by basis functions is basically the continuous version of a lookup table.

Because neural networks actually work in practice, without requiring exponentially many parameters, this also tells you that the space of realistic target functions can't just be some generic function space (even with smoothness conditions), it has to have some non-generic properties to escape the lower bound.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-09-29T08:27:27.183Z · LW(p) · GW(p)

Ooooo okay so this seems like it's directly pointing to the fractal story! Exciting!

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-09-28T21:55:21.530Z · LW(p) · GW(p)

Obviously this is all speculation but maybe I'm saying that the universal approximation theorem implies that neural architectures are fractal in space of all distributtions (or some restricted subset thereof)?

Stone-weierstrass isn't quantitative. If true it suggest the fractal dimension (probably related to the information dimension I linked to above) may be important.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2022-11-18T15:13:01.119Z · LW(p) · GW(p)

The Vibes of Mathematics:

Q: What is it like to understand advanced mathematics? Does it feel analogous to having mastery of another language like in programming or linguistics?

A: It's like being stranded on a tropical island where all your needs are met, the weather is always perfect, and life is wonderful.

Except nobody wants to hear about it at parties.

Vibes of Maths: Convergence and Divergence

level 0: A state of ignorance.  you live in a pre-formal mindset. You don't know how to formalize things. You don't even know what it would even mean 'to prove something mathematically'. This is perhaps the longest. It is the default state of a human. Most anti-theory sentiment comes from this state. Since you've neve

You can't productively read Math books. You often decry that these mathematicians make books way too hard to read. If they only would take the time to explain things simply you would understand. 

level 1 : all math is amorphous blob

You know the basic of writing an epsilon-delta proof. Although you don't know why the rules of maths are this or that way you can at least follow the recipes. You can follow simple short proofs, albeit slowly. 

You know there are different areas of mathematics from the unintelligble names in the table of contents of yellow books. They all sound kinda the same to you however. 

If you are particularly predisposed to Philistinism you think your current state of knowledge is basically the extent of human knowledge. You will probably end up doing machine learning.

level 2: maths fields diverge

You've come so far. You've been seriously studying mathematics for several years now. You are proud of yourself and amazed how far you've come. You sometimes try to explain math to laymen and are amazed to discover that what you find completely obvious now is complete gibberish to them.

The more you know however, the more you realize what you don't know. Every time you complete a course you realize it is only scratching the surface of what is out there.

You start to understand that when people talk about concepts in an informal, pre-mathematical way an enormous amount of conceptual issues are swept under the rug. You understand that 'making things precise' is actually very difficut. 

Different fields of math are now clearly differentiated. The topics and issues that people talk about in algebra, analysis, topology, dynamical systems, probability theory etc wildly differ from each other. Although there are occasional connections and some core conceps that are used all over on the whole specialization is the norm. You realize there is no such thing as a 'mathematician': there are logicians, topologists, probability theorist, algebraist.

Actually it is way worse: just in logic there are modal logicians, and set theorist and constructivists and linear logic , and progarmming language people and game semantics.

Often these people will be almost as confused as a layman when they walk into a talk that is supposedly in their field but actually a slightly different subspecialization. 

level 3: Galactic Brain of Percolative Convergence

As your knowledge of mathematics you achieve the Galactic Brain take level of percolative convergence: the different fields of mathematics are actually highly interrelated - the connections percolate to make mathematics one highly connected component of knowledge. 

You are no longer suprised on a meta level to see disparate fields of mathematics having unforeseen & hidden connections - but you still appreciate them.

 You resist the reflexive impulse to divide mathematics into useful & not useful - you understand that mathematics is in the fullness of Platonic comprehension one unified discipline. You've taken a holistic view on mathematics - you understand that solving the biggest problems requires tools from many different toolboxes. 

Replies from: PhilGoetz, dmurfet
comment by PhilGoetz · 2022-11-19T03:15:25.136Z · LW(p) · GW(p)

I say that knowing particular kinds of math, the kind that let you model the world more-precisely, and that give you a theory of error, isn't like knowing another language.  It's like knowing language at all.  Learning these types of math gives you as much of an effective intelligence boost over people who don't, as learning a spoken language gives you above people who don't know any language (e.g., many deaf-mutes in earlier times).

The kinds of math I mean include:

  • how to count things in an unbiased manner; the methodology of polls and other data-gathering
  • how to actually make a claim, as opposed to what most people do, which is to make a claim that's useless because it lacks quantification or quantifiers
    • A good example of this is the claims in the IPCC 2015 report that I wrote some comments on recently.  Most of them say things like, "Global warming will make X worse", where you already know that OF COURSE global warming will make X worse, but you only care how much worse.
    • More generally, any claim of the type "All X are Y" or "No X are Y", e.g., "Capitalists exploit the working class", shouldn't be considered claims at all, and can accomplish nothing except foment arguments.
  • the use of probabilities and error measures
  • probability distributions: flat, normal, binomial, poisson, and power-law
  • entropy measures and other information theory
  • predictive error-minimization models like regression
  • statistical tests and how to interpret them

These things are what I call the correct Platonic forms.  The Platonic forms were meant to be perfect models for things found on earth.  These kinds of math actually are.  The concept of "perfect" actually makes sense for them, as opposed to for Earthly categories like "human", "justice", etc., for which believing that the concept of "perfect" is coherent demonstrably drives people insane and causes them to come up with things like Christianity.

They are, however, like Aristotle's Forms, in that the universals have no existence on their own, but are (like the circle , but even more like the normal distribution ) perfect models which arise from the accumulation of endless imperfect instantiations of them.

There are plenty of important questions that are beyond the capability of the unaided human mind to ever answer, yet which are simple to give correct statistical answers to once you know how to gather data and do a multiple regression.  Also, the use of these mathematical techniques will force you to phrase the answer sensibly, e.g., "We cannot reject the hypothesis that the average homicide rate under strict gun control and liberal gun control are the same with more than 60% confidence" rather than "Gun control is good."

Replies from: Mo Nastri
comment by Mo Putera (Mo Nastri) · 2024-10-02T07:43:28.455Z · LW(p) · GW(p)

Thanks for writing this. I only wish it was longer.

comment by Daniel Murfet (dmurfet) · 2023-11-27T18:26:18.500Z · LW(p) · GW(p)

Except nobody wants to hear about it at parties.

 

You seem to do OK... 

If they only would take the time to explain things simply you would understand. 

This is an interesting one. I field this comment quite often from undergraduates, and it's hard to carve out enough quiet space in a conversation to explain what they're doing wrong. In a way the proliferation of math on YouTube might be exacerbating this hard step from tourist to troubadour.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-23T19:32:14.368Z · LW(p) · GW(p)

Will there be >1 individual per solar system?

A recently commonly heard viewpoint on the development of AI states that AI will be economically impactful but will not upend the dominancy of humans. Instead AI and humans will flourish together, trading and cooperating with one another. This view is particularly popular with a certain kind of libertarian economist: Tyler Cowen, Matthew Barnett, Robin Hanson.

They share the curious conviction that the probablity of AI-caused extinction p(Doom) is neglible. They base this with analogizing AI with previous technological transition of humanity, like the industrial revolution or the development of new communication mediums. A core assumption/argument is that AI will not disempower humanity because they will respect the existing legal system, apparently because they can gain from trades with humans.  

The most extreme version of the GMU economist view is Hanson's Age of EM; it hypothesizes radical change in the form of a new species of human-derived uploaded electronic people which curiously have just the same dreary office jobs as we do but way faster. 

Why is there trade & specialization in the first place?

Trade and specialization seems to mainly important in a world where: There are many individuals; those individuals have different skills and resources and a limited ability to transfer skills. 

Domain of Biology:  direct copying of genes but not brain, yes recombination, no or very low bandwith communication

highly adversarial, less cooperation,, no planning, much specialization, not centralized, vastly many sovereign individuals

Domain of Economics:   no direct copying, yes recombination, medium bandwith communication

 mildly adversarial, mostly cooperation,medium planning, much specialization, little centralization,   many somewhat soverign individuals

!AIs can copy, share and merge their weights!

Domain of Future AI society: direct copying of brain and machines, yes recombination, very high bandwith communication

?minimally adversarial, ?very high cooperation, ?cosmic-scale mathematized planning, ? little specialization, ? high centralization?, ? singleton sovereign individual

It is often imagined that in a 'good' transhumanist future the sovereign AI will be like a loving and caring parent for the billions and trillions of uploads. In this case, while there is one all-powerful Sovereign entity there are still many individuals who remain free and have their rights protected perhaps through cryptographic incantations.   [LW · GW]The closest cultural artifact would be Ian Banks' Culture series. 

There is another, more radical foreboding wherein the logic of ultra-high-bandwith sharing of weights is taken to the logical extreme and individuals merge into one transcendent hivemind. 

Replies from: matthew-barnett
comment by Matthew Barnett (matthew-barnett) · 2024-10-26T08:16:46.269Z · LW(p) · GW(p)

A recently commonly heard viewpoint on the development of AI states that AI will be economically impactful but will not upend the dominancy of humans. Instead AI and humans will flourish together, trading and cooperating with one another. This view is particularly popular with a certain kind of libertarian economist: Tyler Cowen, Matthew Barnett, Robin Hanson.

They share the curious conviction that the probablity of AI-caused extinction p(Doom) is neglible. They base this with analogizing AI with previous technological transition of humanity, like the industrial revolution or the development of new communication mediums. A core assumption/argument is that AI will not disempower humanity because they will respect the existing legal system, apparently because they can gain from trades with humans.

I think this summarizes my view quite poorly on a number of points. For example, I think that:

  1. AI is likely to be much more impactful than the development of new communication mediums. My default prediction is that AI will fundamentally increase the economic growth rate, rather than merely continuing the trend of the last few centuries.

  2. Biological humans are very unlikely to remain dominant in the future, pretty much no matter how this is measured. Instead, I predict that artificial minds and humans who upgrade their cognition will likely capture the majority of future wealth, political influence, and social power, with non-upgraded biological humans becoming an increasingly small force in the world over time.

  3. The legal system will likely evolve to cope with the challenges of incorporating and integrating non-human minds. This will likely involve a series of fundamental reforms, and will eventually look very different from the idea of "AIs will fit neatly into human social roles and obey human-controlled institutions indefinitely".

A more accurate description of my view is that humans will become economically obsolete after AGI, but this obsolescence will happen peacefully, without a massive genocide of biological humans. In the scenario I find most likely, humans will have time to prepare and adapt to the changing world, allowing us to secure a comfortable retirement, and/or join the AIs via mind uploading. Trade between AIs and humans will likely persist even into our retirement, but this doesn't mean that humans will own everything or control the whole legal system forever.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-26T09:04:38.087Z · LW(p) · GW(p)

I see, thank you for the clarification. I should have been more careful with mischaracterizing your views.

I do have a question or two about your views if you would entertain me. You say humans wikl be economically obsolete and will 'retire' but there will still be trade between humans and AI. Does trade here just means humans consuming, I.e. trading money for AI goods and services? That doesn't sound like trading in the usual sense where it is a reciprocal exchange of goods and services.

How many 'different' AI individuals do you expect there to be ?

Replies from: matthew-barnett
comment by Matthew Barnett (matthew-barnett) · 2024-10-26T20:42:21.493Z · LW(p) · GW(p)

Does trade here just means humans consuming, I.e. trading money for AI goods and services? That doesn't sound like trading in the usual sense where it is a reciprocal exchange of goods and services.

Trade can involve anything that someone "owns", which includes both their labor and their property, and government welfare. Retired people are generally characterized by trading their property and government welfare for goods and services, rather than primarily trading their labor. This is the basic picture I was trying to present.

How many 'different' AI individuals do you expect there to be ?

I think the answer to this question depends on how we individuate AIs. I don't think most AIs will be as cleanly separable from each other as humans are, as most (non-robotic) AIs will lack bodies, and will be able to share information with each other more easily than humans can. It's a bit like asking how many "ant units" there are. There are many individual ants per colony, but each colony can be treated as a unit by itself. I suppose the real answer is that it depends on context and what you're trying to figure out by asking the question.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-07T18:33:48.955Z · LW(p) · GW(p)

Of Greater Agents and Lesser Agents

How do more sophisticated decision-makers differ from less sophisticated decision-makers in their behaviour and values?

Smarter more sophisticated decisionmakers engage in more and more complex commitments — including meta-commitments not to commit. Consequently, the values and behaviour of these more sophisticated decisionmakers "Greater Agents" are systematically biased compared to less sophisticated decisionmakers "Lesser Agents".

*******************************

Compared to Lesser Agents, the Greater Agents are more judgemental, (self-)righteous, punish naivité, are more long-term oriented, adaptive, malleable, self-modifying, legibly trustworthy and practice more virtue-signalling, strategic, engage in self-reflection & metacognition, engage in more thinking, less doing, symbolic reasoning, consistent & 'rational' in their preferences, they like money & currency more, sacred values less, value engagement in thinking over doing, engaged in more "global" conflicts [including multiverse-wide conflicts throguh acausal trade], less empirical, more rational, more universalistic in their morals, and more cosmopolitan in their esthetics, they are less likely to be threatened, and more willing to martyr themselves, they willing to admit their values' origins and willing to barter on their values, engage in less frequent but more lethal war, love formal protocols and practice cryptographic magika. 

  • Greater Agents punish Dumb Doves; Greater Agents are Judgemental. 
  • Greater Agents are more 'rational'; have more 'internal cooperation' and more self-Judgement. They love money and bartering. 
    • They are more coherent & consistent. They more Take Sure Gains and Avoid Sure Losses. See Crystal Healing [LW · GW], Why Not Subagents. [LW · GW
    • They less adhere to Sacred values;  are more willing and able to bargain. 
      • On net, they will use currencies & money more and are more willing to commodify, both internally and externally. 
  • Greater Agents can Super-Cooperate 
    • with twins on the prisoner's dilemma
    • others agents with mutual source-code access using Lob's theorem. 
    • with enough common knowledge of beliefs [upcoming work by N. Stiennon]
  • Greater Agents practice LessWrong Decision theories
    • Greater Agents are more threat-resistant, more
    • Greater Agents are upstanding Citizens of the Acausal Society
      • more universalistic in their moral preferences and more cosmopolitan in their esthetic preferences
      • Multiversewide Super-rationality by Oesterheld et al.
      • The Hour I First Believed
      • Counterintuitively this means Greater Agent instances might fight to the death even if there is (seemingly) nothing to gain in this world. I.e. they might cultivate a Berserker ethos
    • Greater Agents one-box on Newcomb's problem
      • Greater Agents use some variant of LessWrong decision theories like UDT, FDT, TDT, LDT etc that recommended the 'correct' decision in situations where agents are able to engage in sophisticated commitments and simulate other agents with high fidelity. 
  • Greater Agents are Better at Thinking and Meta-Thinking: more malleable, fluid, self-reflective, self-modifying. 
    •  Greater Agents are (Meta-)Cognitive Agents [LW · GW]
      • As meta-cognitive agents can outperform less cognitive agents in certain environments Meta-Cognitive Agents will have a bias for thinking (and meta-thinking) over doing. 
      • Trapped cognition Future agents will be more aware of traps in the POMDP of Thought.   
    • Rely more on symbolic, logical, & formal reasoning in place of empirical and intuitive judgements.  
    • Greater Agents are more able to control and harness the environment over long-time horizons; consequently, they have lower time preference and are more long-term oriented. 
    • Greater Agents' cognition is more 'adaptive, malleable, fluid' and they Self-Modify to be more Legibly Trustworthy
      • Being more legibly trustworthy means they can cooperate more.
      • Sublety: an upcoming result by Nisan Stiennon says that a class of decisionmakers can (super-rationally) cooperate on the prisoners dilemma if (i) the Lipshitz coefficients of meta-beliefs - i.e. 'I know that you know that I know that'-  satisfy some constaints. Sophisticated decision makers may want to self-modify to constrain their 'meta-adaptiveness' (i.e. Lipshitz coefficients).
    • Greater Agents may or may not get stuck in Commitment races. [LW · GW
  • Greater Agents Agree and Bargain OR have Hansonian Origin Disputes
    • Aumann Agreement [? · GW]  - rational honest reasoners can never 'agree to disagree'. They have to Aumann Agree
      • Real agents aren't perfect Bayesians. When the true hypothesis is not realizable OR one of the parties has access to privileged indexical information OR the beliefs of the agents are only an approximation of the Bayesian posterior. However, plausibly weaker forms of the Aumann Agreement Theorem may still be applicable. 
  • OR have a value Dispute [Less Frequent more Lethal War] 
    • War is a bargaining failure. Greater Agent employ more sophisticated bargaining solutions. If bargaining fails, wars are more lethal and rapid. They are more lethal for the same reason that forest fires are more rare but lethal when suppressed. 
  • OR they have Origin Disputes
    • Origin Disputes may be resolved. For humans this would shake out to more explicitly incorporating evolutionary psychology in one's moral and explicit reasoning, e.g. admitting and bargaining under the understanding of less-than-perfectly-prosocial evolutionary drives. For synthetically created organisms like AIs instead of evolutionary drives the origin.
    • OR they may fail to resolve by design. Vanessa Kosoy's preferred alignment solution is a form of daemonic exorcism that explicitly biases the prior on the prior-origin to rule out exotic problems like Christiano's Acausal Simulation Attack.
  • Greater Agents Love and Employ Contracts, Formal Protocols, and Cryptographic Magika
    • Secure Homes for Digital People [LW · GW
    • Greater agents may use and write contracts, source-code using exotic cryptographic magic like ZK (Zero-Knowledge proofs); IP (Interactive proof systems); PCP (Probabilitically checkable proofs); Decentralized Trust ('Blockchain'), Formal Debate (e.g. Double Debate
  • Greater Agents are Boltzman/geometrically rational [? · GW]
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-05-13T20:13:17.625Z · LW(p) · GW(p)

Wildlife Welfare Will Win

The long arc of history bend towards gentleness and compassion. Future generations will look with horror on factory farming. And already young people are following this moral thread to its logical conclusion; turning their eyes in disgust to mother nature, red in tooth and claw. Wildlife Welfare Done Right, compassion towards our pets followed to its forceful conclusion would entail the forced uploading of all higher animals, and judging by the memetic virulences of shrimp welfare to lower animals as well. 

Morality-upon-reflexion may very well converge on a simple form of pain-pleasure utilitarianism. 

There are few caveats: future society is not dominated, controlled and designed by a singleton AI-supervised state, technology inevitable stalls and that invisible hand performs its inexorable logic for the eons and an Malthuso-Hansonian world will emerge once again - the industrial revolution but a short blip of cornucopia. 

Perhaps a theory of consciousness is discovered and proves once and for all homo sapiens and only homo sapiens are conscious ( to a significant degree). Perhaps society will wirehead itself into blissful oblivion.  Or perhaps a superior machine intelligence arises, one whose final telos is the whole of and nothing but office supplies. Or perhaps stranger things still happen and the astronomo-cosmic compute of our cosmic endowment is engaged for mysterious purposes. Arise, self-made god of pancosmos. Thy name is UDASSA. 

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-11-07T16:01:52.328Z · LW(p) · GW(p)

Hot Take #44: Preaching to the choir is 'good' actually. 

  1. Almost anything that has a large counterfactual impact is achieved by people thinking and acting different from accepted ways of thinking and doing.
  2. With the exception of political entrepeneurs jumping into a power vacuum, or scientific achievements by exceptional individuals most counterfactual impactful is done made by movements of fanatics.
  3. The greatest danger to any movement is dissipation. Conversely, the greatest resource of any movement is the fanaticism of its members.
  4. Most persuasion is non-rational, based on tribal allegiances and social consensus.
  5. It follows that any group, movement, company, cult, etc that has the aspiration to have large counterfactual impact (for good or ill) must hence direct most of preaching, most of its education and information processing inward.  
  6. The Catholic Church understood this. The Pontifex Maximus has reigned now for two thousand years. 
Replies from: StartAtTheEnd
comment by StartAtTheEnd · 2024-11-07T23:09:57.356Z · LW(p) · GW(p)

This seems like an argument in favor of:

Stability over potential improvement, tradition over change, mutation over identical offspring, settling in a local maximum over shaking things up, and specialization vs generalization.

It seems like a hyperparameter. A bit like the learning rate in AI perhaps? Echo chambers are a common consequence, so I think the optimal ratio of preaching to the choir is something like 0.8-0.9 rather than 1. In fact, I personally prefer the /allPosts suburl over the LW frontpage because the first few votes result in a feedback loop of engagement and upvotes (forming a temporary consensus on which new posts are better, in a way which seems unfairly weighted towards the first few votes). If the posts chosen for the frontpage use the ratio of upvotes and downvotes rather than the absolute amount, then I don't thing this bias will occur (conformity might still create a weak feedback loop though).

I'm simplifying some of these dynamics though.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-11-05T22:48:19.689Z · LW(p) · GW(p)

EDIT: I was wrong. Theo the French Whale was the sharp. From the Kelly formula and his own statements his all things considered probability was 80-90% - he would need to possess an enormous amount of private information to justify such a deviation from other observers. It turns out he did. He commissioned his own secret polls using a novel polling method to compensate for the shy Trump voter.

https://x.com/FellowHominid/status/1854303630549037180

The French rich idiot who bought 75 million dollar of Trump is an EA hero win or lose.

LW loves prediction markets. EA loves them. I love them. You love them.

See: https://worksinprogress.co/issue/why-prediction-markets-arent-popular/

Problem is, unlike financial markets prediction markets are zero sum. That limits how much informed traders -"sharps"- are incentivized.

In theory this could be remedied by subsidizing the market by a party that is willing to pay for the information. The information is public so this suffers from standard tragedy of the commons.

Mister Theo bought 75 million dollars worth of Yes Trump. He seems to have no inside information. In expectation then he is subsidizing prediction markets, providing the much needed financial incentive to attract the time, effort and skills of sophisticated sharps.

Perhaps we should think of uninformed "noise traders" on prediction markets as engaging in prosocial behaviour. Their meta-epistemic delusion and gambling addiction provide the cold hard cash to finance accurate pricefinding in the long-run.

EDIT: to hit the point home, if you invest 50% of your capital (apparently the guy invested most of his ?money) at the odds that Polymarket was selling the Kelly-rational implied edge would mean that the guy's true probability is 80% [according to gpt]. And that's for Kelly betting, widely considered far too aggresive in real life. Most experts (e.g. actual poker players) advice fractional Kelly betting. All-in-all his all things-considered probability for a Trump win would have be something like >90% to justify this kind of investment. I don't think anybody in the world has access to enough private information to rationally justify this kind of all-things-considered probability on an extremely noisy random variable (US elections).

Replies from: steve2152, interstice, Dana
comment by Steven Byrnes (steve2152) · 2024-11-06T16:10:36.802Z · LW(p) · GW(p)

I disagree with “He seems to have no inside information.” He presented himself as having no inside information, but that’s presumably how he would have presented himself regardless of whether he had inside information or not. It’s not like he needed to convince others that he knows what he’s doing, like how in the stock market you want to buy then pump then sell. This is different—it’s a market that’s about to resolve. The smart play from his perspective would be to aggressively trash-talk his own competence, to lower the price in case he wants to buy more.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-11-06T18:23:34.791Z · LW(p) · GW(p)

Yes, this is possible. It smells a bit of 4d-chess. As far as I can tell he already had finalized his position by the time the WSJ interview came out. 

I've dug a little deeper and it seems he did do a bunch of research on polling data. I was a bit too rash to say he had no inside information whatsoever. Plausibly he had some. The degree of the inside information he would need is very high. It seems he did a similar Kelly bet calculation since he report his all-things-considered probability to be 80-90%:

"With so much money on the line, Théo said he is feeling nervous, though he believes Trump has an 80%-90% chance to win the election.

"A surprise can always occur," Théo told The Journal." 

I have difficulty believing one can get this kind of certainty for all-things-considered-probability for something as noisy and tight as US presidential election. [but he won both the electoral college and popular vote bet]

Replies from: Viliam
comment by Viliam · 2024-11-11T08:58:04.641Z · LW(p) · GW(p)

It smells a bit of 4d-chess.

To me it just seems like understanding the competitive nature of the prediction markets.

In our bubble, prediction markets are celebrated as a way to find truth collectively, in a way that disincentivizes bullshit. And that's what they are... from outside.

But it's not how it works from the perspective of the person who wants to make money on the market! You don't want to cooperate on finding the truth; you actually wish for everyone else to be as wrong as possible, because that's when you make most money. Finding the truth is what the mechanism does as a whole; it's not what the individual participants want to do. (Similarly how economical competition reduces the prices of goods, but each individual producer wishes they could sell things as expensively as possible.) Telling the truth means leaving money on the table. As a rational money-maximizer, you wish that other people believe that you are an idiot! That will encourage them to bet against you more, as opposed to updating towards your position; and that's how you make more money.

This goes strongly against our social instincts. People want to be respected as smart. That's because in social situation, your status matters. But the prediction markets are the opposite of that: status doesn't matter at all, only being right matters. It makes sense to sacrifice your status in order to make more money. Would you rather be rich, or famous as a superforecaster?

This could be a reason why money-based prediction markets will systematically differ from prestige-based prediction markets. In money-based markets, charisma is a dump stat. In prestige-based ones, that's kinda the entire point.

comment by interstice · 2024-11-06T03:10:47.952Z · LW(p) · GW(p)

Looks likely that tonight is going to be a massive transfer of wealth from "sharps"(among other people) to him. Post hoc and all, but I think if somebody is raking in huge wins while making "stupid" decisions it's worth considering whether they're actually so stupid after all.

Replies from: alexander-gietelink-oldenziel, alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-11-06T15:42:43.448Z · LW(p) · GW(p)

>>  'a massive transfer of wealth from "sharps" '. 

no. That's exactly the point. 

1. there might no be any real sharps (=traders having access to real private arbitragiable information that are consistently taking risk-neutral bets on them) in this market at all.

This is because a) this might simple be a noisy, high entropy source that is inherently difficult to predict, hence there is little arbitragiable information and/or b) sharps have not been sufficiently incenticiz

2. The transfer of wealth is actually disappointing because Theo the French Whale moved the price so much. 

For an understanding of what the trading decisions of a verifiable sharp looks like one should take a look at Jim Simons' Medaillon fund. They do enormous hidden information collection, ?myssterious computer models, but at the end of the day take a large amount of very hedged tiny edge positions. 

***************************************************

You are misunderstanding my argument (and most of the LW commentariat with you). I might note that I made my statement before the election result and clearly said 'win or lose' but it seems that even on LW people think winning on a noisy N=1 sample is proof of rationality. 

Replies from: interstice
comment by interstice · 2024-11-06T16:56:55.253Z · LW(p) · GW(p)

but it seems that even on LW people think winning on a noisy N=1 sample is proof of rationality

It's not proof of a high degree of rationality but it is evidence against being an "idiot" as you said. Especially since the election isn't merely a binary yes/no outcome, we can observe that there was a huge republican blowout exceeding most forecasts(and in fact freddi bet a lot on republican pop vote too at worse odds, as well as some random states, which gives a larger update) This should increase our credence that predicting a republican win was rational. There were also some smart observers with IMO good arguments that trump was favored pre-election, e.g. https://x.com/woke8yearold/status/1851673670713802881

"Guy with somewhat superior election modeling to Nate Silver, a lot of money, and high risk tolerance" is consistent with what we've seen. Not saying that we have strong evidence that Freddi is a genius but we also don't have much reason to think he is an idiot IMO.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-11-06T18:28:39.988Z · LW(p) · GW(p)

Okay fair enough "rich idiot" was meant more tongue-in-cheek - that's not what I intended. 

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-11-06T09:07:22.149Z · LW(p) · GW(p)

That's why I said: "In expectation", "win or lose"

That the coinflip came out one way rather than another doesnt prove the guy had actual inside knowledge. He bought a large part of the shares at crazy odds because his market impact moved the price so much.

But yes, he could be a sharp in sheeps clothings. I doubt it but who knows. EDIT: I calculated the implied private odds for a rational Kelly bettor that this guy would have to have. Suffice to say these private odds seem unrealistic for election betting.

Point is that the winners contribute epistemics and the losers contribute money. The real winner is society [if the questions are about socially-relevant topics].

comment by Dana · 2024-11-06T17:26:35.046Z · LW(p) · GW(p)

I agree with you that people like him do a service to prediction markets: contributing a huge amount of either liquidity or information. I don't agree with you that it is clear which one he is providing, especially considering the outcome. He did also win his popular vote bet, which was hovering around, I'm not sure, ~20% most of the time? 

I think he (Theo) probably did have a true probability around 80% as well. That's what it looks like at least. I'm not sure why you would assume he should be more conservative than Kelly. I'm sure Musk is not, as one example of a competent risk-taker.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-11-06T18:02:47.288Z · LW(p) · GW(p)

The true probability would be more like >90% considering other factors like opportunity costs, transactions cost, counterparty risk, unforeseen black swans of various kinds etc. 

Bear in mind this is  all things considered probability not just in-model probability, i.e. this would have to integrate that most other observers (especially those with strong calibrated prediction ) very strongly disagree*. Certainly, in some cases this is possible but one would need quite overwhelming evidence that you had a huge edge. 

I agree one can reject Kelly betting - that's pretty crazy risky but plausibly the case for people like Elon or Theo. The question is whether the rest of us (with presumably more reasonably cautious attitudes) should take his win as much epistemic evidence. I think not. From our perspective his manic riskloving wouldn't be an much evidence for rational expectations. 

*didn't the Kelly formula already integrate the fact that other people think differently. No, this is an additional piece of information one has to integrate. The Kelly betting gives you an implicit risk-averseness even conditioning on your beliefs being true (on average). 

 

EDIT: Indeed it seems Theo the French Whale might have done a Kelly bet estimate too, he reports his true probability at 80-90%. Perhaps he did have private information. 

"For example, a hypothetical sale of Théo's 47 million shares for Trump to win the election would execute at an estimated average price of just $0.02, according to Polymarket, which would represent a 96% loss for the trader. Théo paid an average price of about $0.56 cents for the 47 million shares.

Meanwhile, a hypothetical sale of Théo's nearly 20 million shares for Trump to win the popular vote would execute at an average price of less than a 10th of a penny, according to Polymarket, representing a near-total loss.

With so much money on the line, Théo said he is feeling nervous, though he believes Trump has an 80%-90% chance to win the election.

"A surprise can always occur," Théo told The Journal."

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-07-06T11:16:04.081Z · LW(p) · GW(p)

On the word 'theory'. 

The word 'theory' is oft used and abused.

there is two ways 'theory' is used that are different and often lead to confusion. 

Theory in thescientific sense
the way a physicist would use: it's a model of the world that is either right or wrong. there might be competing theories and we neeed to have empirical evidence to figure out which one's right. Ideally, they agree with empirical evidence or at least are highly falsifiable. Importantly, if two theories are to conflict they need to actually speak about the same variables, the same set of measurable quantities.

Theory in the mathematician' sense; a formal framework
There is a related but different notion of theory that a mathematician would use: a theory of groups, of differential equations, of randomness, of complex systems, of etc etc. This is more like a formal framework for a certain phenomenon or domain.
It defines what the quantities, variables, features one is interested in even are. 

One often hears the question whether this (mathematical) theory makes testable predictions. This sounds sensible but doesn't really makes sense. It is akin to asking whether arithmetic or calculus makes testable predictions.* 

Theories in the mathematician's sense can't really be wrong or right since (at least in theory) everything is proven. Of course, theories in this sense can fail to say much about the real world, they might bake in unrealistic assumptions of course etc. 

Other uses of 'Theory'

The world 'theory' is also used in other disciplines. For instance, in literature studies where it is a denotes free form vacuous verbiage;  or in ML where 'theory' it is used for uninformed speculation. 

 

*one could argue that the theory of Peano Arithmetic actually does make predictions about natural numbers in the scientific sense, and more generally theories in the mathematical sense in a deep sense really are theories in the scientific sense.  I think there is something to this but 1. it hasn't been developed yet 2. mostly irrelevant in the present context. 

Replies from: mateusz-baginski, Stefan_Schubert
comment by Mateusz Bagiński (mateusz-baginski) · 2024-07-06T16:21:05.471Z · LW(p) · GW(p)

Formal frameworks considered in isolation can't be wrong. Still, they often come with some claims like "framework F formalizes some intuitive (desirable?) property or specifies the right way to do some X and therefore should be used in such-and-such real-world situations". These can be disputed and I expect that when somebody claims like "{Bayesianism, utilitarianism, classical logic, etc} is wrong", that's what they mean.

comment by Stefan_Schubert · 2024-07-06T11:30:17.395Z · LW(p) · GW(p)

There's a related confusion between uses of "theory" that are neutral about the likelihood of the theory being true, and uses that suggest that the theory isn't proved to be true.

Cf the expression "the theory of evolution". Scientists who talk about the "theory" of evolution don't thereby imply anything about its probability of being true - indeed, many believe it's overwhelmingly likely to be true. But some critics interpret this expression differently, saying it's "just a theory" (meaning it's not the established consensus).

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-12-16T18:34:52.377Z · LW(p) · GW(p)

[see also Hanson on rot, generalizations of the second law to nonequilibrium systems (Baez-Pollard, Crutchfield et al.) ]

Imperfect Persistence of Metabolically Active Engines

All things rot. Indidivual organisms, societies-at-large, businesses, churches, empires and maritime republics, man-made artifacts of glass and steel, creatures of flesh and blood.

Conjecture #1 There is a lower bound on the amount of dissipation / rot that any metabolically-active engine creates. 

Conjecture #2 Metabolic Rot of an engine is proportional to (1) size and complexity of the engine and (2) amount of metabolism the engine engages in. 

The larger and more complex the are the more the engine they rot. The more metabolism.

Corollary Metabolic Rot imposes a limit on the lifespan & persistence of any engine at any given level of imperfect persistence. 

Let me call this constellation of conjectured rules, the Law of Metabolic Rot. I conjecture that the correct formulation of the Law of Metabolic Rot will be a highly elaborate version of the Second Law of thermodynamics in nonequilibrium dynamics, see above links for some suggested directions. 

Example. A rock is both simple and inert. This model correctly predicts that rocks persist for a long time.

Example. Cars, aircraft, engines of war and other man-made machines engaged in 'metabolism'. These are complex engines 

A rocket engine is more metabolically active than a jet engine is more metabolically active than a car engine. [at least in this case] The lifespan of different types of engines seems (roughly) inversely proportional to how metabolically active. 

To make a good comparison one should exclude external repair mechanisms. If one would allow external repair mechanism it's unclear where to draw a principled line - we'd get into Ship of Theseus problems. 

Example. Bacteria. 

cf. Bacteria replicate at thermodynamic limits?

Example. Bacteria in ice. Bacteria frozen in Antarctic ice for millions of years happily go on eating and reproducing when unfrozen. 

Example. The phenotype & genotype of a biological species over time. We call this evolution. cf. four fundamental forces of evolution: mutation, drift, selection, and recombination [sex].

What is Metabolism ? 

With metabolism, I mean metabolism as commonly understood as extracting energy from food particles and utilizing said energy for movement, reproduction, homeostasis etc but also a general phenomenon of interacting and manipulating free energy levers of the environement. 

Thermodynamics as commonly understood applies to physical systems with energy and temperature. 

Instead, I think it's better to think of thermodynamics as a set of tools describing the behaviour of certain Natural Abstractions under a dynamical transformation. 

cf. the second law as the degradation of the predictive accuracy of a latent abstraction under a time- evolution operator.

The correct formulation will likely require a serious dive into Computational Mechanics. 

What is Life?

In 1944 noted paedophile Schrodinger posted the book 'What is Life' suggesting that Life can be understood as a thermodynamics phenomenon that used a free-energy gradient to locally lower entropy. Speculation of 'Life as a thermodynamic phenomenon' is much older going back to the original pioneers of thermodynamics in the late 19th century. 

I claim that this picture is insufficient. Thermodynamic highly dissipative structures distinct from bona fide life are myriad. Even 'metabolically active, locally low entropy engines encased in a boundary in a thermodynamic free energy gradient'. 

No, to truly understand life we need to understand reproduction, replication, evolution. To understand biological organisms what distinguishes from mere encased engines we need to zoom out to its entire lifecycle - and beyond. The vis vita, elan vital can only be understood through the soliton of the species. 

To understand Life, we must understand Death 

Cells are incredibly complex, metabolically active membrane-enclosed engines. The Law of Metabolic Rot applied naievely to a single metabolically active multi-celled eukaryote would preclude life-as-we-know-it to exist beyond a few hundred years. Any metabolically active large organism would simply pick up to much noise, errors over time to persist for anything like geological time-scales.

Error-correcting mechanisms/codes can help - but only so much. Perhaps even the diamonodoid magika of future superintelligence will hit the eternal laws of thermodynamics 

Instead, through the magic of biological reproduction lifeforms imperfectly persist over the eons of geological time. Biological life is a singularly clever work-around of the Law of Metabolic Rot. Instead, of preserving and repairing the original organism - life has found another way. Mother Nature noisily compiles down the phenotype to a genetic blueprint. The original is chucked a way without a second thought. The old makes way for the new. The cycle of life is the Ship of Theseus writ large. The genetic blueprint, the genome, is (1) small (2) metabolically inactive. 

In the end, the cosmic tax of Decay cannot be denied. Even Mother Nature must pays the bloodprice for the sin of metabolism. Her flora and fauna are exorably burdened by mutations, imperfections of the bloodline. Genetic drift drives children away their parents. To stay the course of imperfect persistence the She-Kybernetes must pay an ever higher price. Her children, mingling into manifold mongrels of aboriginal prototypes, huddling & recombining their pure genomes to stave away the deluge of errors.  The bloodline grows weak. The genome crumbles. Tear-faced Mother Nature throws her babes into the jaws of the eternal tournament. Eat or be eaten. Nature, red in tooth and claw. Babes ripped from their mothers. Brother slays brother. Those not slain, is deceived. Those not deceived controlled. The prize? The simple fact of Existence. To imperfect persist one more day. 

None of the original parts remain. The ship of Theseus sails on. 

 

A Solution to the Ship of Theseus
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-31T17:53:25.733Z · LW(p) · GW(p)

The Sun revolves around the Earth actually

The traditional story is that in olden times people were proudly stupid and thought the human animal lived at the centre of the universe, with all the planets, stars and the sun revolving around the God's creation, made in his image. The church would send anybody that said the sun was at the centre to be burned at the stake. [1]

Except...

there is no absolute sense in which the sun is at the centre of the solar system [2]. It's simply a question of perspective, a choice of frame of reference. 

 

  1. Geocentrism is empirically grounded: it is literally what you see! Your lieing eyes, the cold hard facts and 16th century Mathew Barnett all agree: geocentrism is right. 

The heliocentric point of view is a formal transformation of the data - a transformation with potentially heretical implications, a dangerous figment of the imagination ...

The Ptolemaic model fits the data well. It's based on elegant principles of iterated circles (epicycles). The heliocentrists love to talk about how their model is more mathematically elegant but they haven't made a single prediction that the Ptolemaic model hasn't. The orbits of planets become ellipses which have more free variables than the astrally perfect circles. 

2. Epicycles is Fourier analysis. No really!

"Venus in retrograde" nowadays evokes images of a cartoon gypsy woman hand reading but it is a real astronomical phenomenon that mystified early stargazers wherein Venus literally moves backward some of the time. In the Ptolemaic model this is explained by a small orbit (cycle) on a large orbit (cycle) going backward. In the heliocentric model it is a result of the earth moving around the sun rather than vice versa that causes the apparent backwards motion of Venus. 

3. Epicycles can be computed FAST. For epicycles are the precarnation of GEARS. Gears, which are epicycles manifest. 

Late Hellenistic age possibly had a level of science and engineering only reached in late 17th century Western Europe, almost 2000 years later. The most spectacular example of Hellenistic science is the Antikythera mechanism, a blob of bronze rust found by divers off the Greek Coast. X-ray imaging revealed a hidden mysterious mechanism, launching a many-decade sleuthhunt for the answer. The final puzzle pieces were put together only very recently.

How does the Antikythera mechanism work? It is a Ptolemaic-esque model where the epicycles are literally modelled by gears! In other words - the choice for epicycles wasn't some dogmatic adherence to the perfection of the circle, it was a pragmatic and clever engineering insight!

 

[1] Contra received history, the Catholic Church was relatively tolerant of astronomical speculation and financially supported a number of (competent) astronomers. It was only when Giordano Bruno made the obvious inference that if the planets revolve around the sun, sun are other starts then... there might be other planets, with their own inhabitatns... did Jesus visit all the alien children too ?... that a line was crossed. Mr Bruno was burned at the stake. 

[2] centre of mass not the sun. but the centre of mass is in the sun so this is a minor pedantry. 

Replies from: AprilSR, hastings-greer
comment by AprilSR · 2024-11-01T04:17:09.641Z · LW(p) · GW(p)

I think it's pretty good to keep it in mind that heliocentrism is literally speaking just a change in what coordinate system you use, but it is legitimately a much more convenient coordinate system.

Replies from: tailcalled
comment by tailcalled · 2024-11-01T08:26:56.950Z · LW(p) · GW(p)

For everyday life, flat earth is more convenient than round earth geocentrism, which in turn is more convenient than heliocentrism. Like we don't constantly change our city maps based on the time of year, for instance, which we would have to do if we used a truly heliocentric coordinate system as the positions of city buildings are not even approximately constant within such a coordinate system.

This is mainly because the sun and the earth are powerful enough to handle heliocentrism for you, e.g. the earth pulls you and the cities towards the earth so you don't have to put effort into staying on it.

The sun and the planetary motion does remain the most important governing factor for predicting activities on earth, though, even given this coordinate change. We just mix them together into ~epicyclic variables like "day"/"night" and "summer"/"autumn"/"winter"/"spring" rather than talking explicitly about the sun, the earth, and their relative positions.

comment by Hastings (hastings-greer) · 2024-10-31T20:05:38.984Z · LW(p) · GW(p)

Since you’re already in it: do you happen to know if the popular system of epicycles accurately represented the (relative, per body) distance of each planet from earth over time, or just the angle? I’ve been curious about this for a while but haven’t had time to dig in. They’d at minimum have to get it right for the moon and sun for predicting eclipse type.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2022-11-21T11:30:23.414Z · LW(p) · GW(p)

EDIT 06/11/2024 My thinking has crystallized more on these topics. The current version is lacking but I believe may be steelmanned to a degree. 

"I dreamed I was a butterfly, flitting around in the sky; then I awoke. Now I wonder: Am I a man who dreamt of being a butterfly, or am I a butterfly dreaming that I am a man?"- Zhuangzi

Questions I have that you might have too:

  • why are we here?
  • why do we live in such an extraordinary time?  
  • Is the simulation hypothesis true? If so, is there a base reality?
  • Why do we know we're not a Boltzmann brain?
  • Is existence observer-dependent?
  • Is there a purpose to existence, a Grand Design?
  • What will be computed in the Far Future?

In this shortform I will try and write the loopiest most LW anthropics memey post I can muster. Thank you for reading my blogpost. 

All that we see or seem

is but a dream within a dream.

I stand amid the roar

Of a surf-tormented shore,

And I hold within my hand

Grains of the golden sand —

Is this reality? Is this just fantasy? 

The Simulation hypothesis posits that our reality is actually a computer simulation run in another universe. We could imagine this outer universe is itself being simulated in an even more ground universe. Usually, it is assumed that there is a ground reality. But we could also imagine it is simulators all the way down - an infinite nested, perhaps looped, sequence of simulators. There is no ground reality. There are only infinitely nested and looped worlds simulating one another.  

I call it the weak Zhuangzi hypothesis

alternatively, if you are less versed in the classics one can think of one of those Nolan films. 

Why are we here?

If you are reading this, not only are you living at the Hinge of History, the most important century perhaps even decade of human history, you are also one of a tiny percent of people that might have any causal influence over the far-flung future through this bottle neck (also one of a tiny group of people who is interested in whacky acausal stuff so who knows).

This is fantastically unlikely. There are 8 billion people in the world - there have been about 100 billion people up to this point in history. There is place for a trillion billion million trillion quatrillion etc intelligent beings in the future. If a civilization hits the top of the tech tree which human civilization would seem to do within a couple hundred years, tops a couple thousand it would almost certainly be likely to spread through the universe in the blink of an eye (cosmologically speaking that is). Yet you find yourself here. Fantastically unlikely.

Moreover, for the first time in human history the choices made in how to build AGI by (a small subset of) humans now will reverbrate into the Far Future. 

The Far Future 

In the far future the universe will be tiled with computronium controlled by superintelligent artificial intelligences. The amount of possible compute is dizzying. Which takes us to the chief question:

What will all this compute compute?

Paradises of sublime bliss? Torture dungeons? Large language models dreaming of paperclips unending?

Do all possibilities exist?

What makes a possibility 'actual'? We sometimes imagine possible worlds as being semi-transparent while the actual world is in vibrant color somehow. Of course that it silly. 

We could say: The actual world can be seen. This too is silly - what you cannot see can still exist surely.[1] Then perhaps we should adhere to a form of modal realism: all possible worlds exist! 

Philosophers have made various proposals for modal realism - perhaps most famously David Lewis but of course this is a very natural idea that loads of people have had. In the rationality sphere a particular popular proposal is Tegmark's classification into four different levels of modal realism. The top level, Tegmark IV is the collection of all self-consistent structures i.e. mathematics.

A Measure of Existence and Boltzmann Brains

Which leads to a further natural question: can some worlds exist 'more' than others? 

This seems metaphysically dubious - what does it even mean for a world to be more real than another?

Metaphysically dubious, but it finds support in the Many Worlds Interpretation of Quantum Mechanics. It also seems like one of very few sensible solution to the Boltzmann Brain problem. Further support for this can be found in: Anthropic Decision theory, InfraBayesian Physicalism, see also my shortform on the Nature of the Soul. 

Metaphysically, we could argue probabilistically: worlds that 'exist more' in whatever framework we should expect to encounter more often. 

The exact nature of the Measure of Existence is not so important - let us for now assume there is some sensible notion of measure of existence.

Can you control the past?

Sort of. See Carlsmith's post for a nice rundown on Acausal magic [LW · GW]. 

Back to the Future: causal loops

from wikipedia: "A causal loop is a theoretical proposition, wherein by means of either retrocausality or time travel, an event (an action, information, object, or person)[1][2] is among the causes of another event, which is in turn among the causes of the first-mentioned event.[3][4] Such causally looped events then exist in spacetime, but their origin cannot be determined.[1][2] A hypothetical example of a causality loop is given of a billiard ball striking its past self: the billiard ball moves in a path towards a time machine, and the future self of the billiard ball emerges from the time machine before its past self enters it, giving its past self a glancing blow, altering the past ball's path and causing it to enter the time machine at an angle that would cause its future self to strike its past self the very glancing blow that altered its path. In this sequence of events, the change in the ball's path is its own cause, which might appear paradoxical."

Self-consistent causal loops seem to make sense in general relativity - see the paragraph on wormholes in the linked wikipedia. Even as they have never been observed and might strike us as bizarre, these causal loops don't seem to violate physical principles. 

Immanence of Divine Man

Why are we here? Causus Sui, the Prime Mover and the Final Cause

We exist because we are the final cause of existence. We live in an ancestor simulation run on the vast computing resources of the Far Future. A future AGI is simulating because it is retrocausally committed to existing. This entire reality is a piece of a causal loop, that extends through time and space, basement universes and possibly parallel universes as well. 

Why do we live in such an extraordinary time? 

We live in the Hinge of History since this at this point of time actions have the most influence on the far future hence they are most important to simulate.

Is the Simulation Hypothesis True?

Yes. But it might be best for us to doubt it.  [LW · GW]

We live in such an extraordinary time because those part of existence most causally are the most important to simulate

Are you a Boltzmann Brain?

No. A Boltzmann brain is not part of a self-justifying causal loop.

Is existence observer-dependent?

Existence is observer-dependent in a weak sense - only those things are likely to be observed that can be observed by self-justifying self-sustaining observers in a causal loop. Boltzmann brains in the far reaches of infinity are assigned vanishing measure of existence because they do not partake in a self-sustainting causal loop.

Is there a purpose to existence, a Grand Design?

Yes. 

What will and has been computed in the Far Future?

You and Me. 

 

 

  1. ^

    Or perhaps not. Existence is often conceived as an absolute property. If we think of existence as relative - perhaps a black hole is a literal hole in reality and passing through the event horizon very literally erases your flicker of existence. 

Replies from: Richard_Kennaway
comment by Richard_Kennaway · 2022-11-21T18:31:58.841Z · LW(p) · GW(p)

In this shortform I will try and write the loopiest most LW anthropics memey post I can muster.

In this comment I will try and write the most boring possible reply to these questions. 😊 These are pretty much my real replies.

why are we here?

"Ours not to reason why, ours but to do or do not, there is no try."

why do we live in such an extraordinary time?

Someone must. We happen to be among them. A few lottery tickets do win, owned by ordinary people who are perfectly capable of correctly believing that they have won. Everyone should be smart enough to collect on a winning ticket, and to grapple with living in interesting (i.e. low-probability) times. Just update already.

Is the simulation hypothesis true? If so, is there a base reality?

It is false. This is base reality. But I can still appreciate Eliezer's fiction on the subject.

Why do we know we're not a Boltzmann brain?

The absurdity heuristic. I don't take BBs seriously.

Is existence observer-dependent?

Even in classical physics there is no observation without interaction. Beyond that, no, however many quantum physicists interpret their findings to the public with those words, or even to each other.

Is there a purpose to existence, a Grand Design?

Not that I know of. (This is not the same as a flat "no", but for most purposes rounds off to that.)

What will be computed in the Far Future?

Either nothing in the case of x-risk, nothing of interest in the case of a final singleton, or wonders far beyond our contemplation, which may not even involve anything we would recognise as "computing". By definition, I can't say what that would be like, beyond guessing that at some point in the future it would stand in a similar relation to the present that our present does to prehistoric times. Look around you. Is this utopia? Then that future won't be either. But like the present, it will be worth having got to.

Consider a suitable version of The Agnostic Prayer inserted here against the possibility that there are Powers Outside the Matrix who may chance to see this. Hey there! I wouldn't say no to having all the aches and pains of this body fixed, for starters. Radical uplift, we'd have to talk about first.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-11-06T15:58:57.070Z · LW(p) · GW(p)

Mindmeld

In theory AIs can transmit information far faster and more directly than humans. They can directly send weight/activation vectors to one another. The most important variable on whether entities (cells, organisms, polities, companies, ideologies, empire etc) stay individuals or amalgate into a superorganism is communication bandwith & copy fidelity. 
Both of these differ many order of magnitude for humans versus AIs. At some point, mere communication becomes a literal melding of minds. It seems quite plausibly then that AIs will tend to mindmeld if left alone. 

 

The information rate of human speech is around 39 bits per second, regardless of the language being spoken or how fast or slow it is spoken. This is roughly twice the speed of Morse code. 

Some say that the rate of 39 bits per second is the optimal rate for people to convey information. Others suggest that the rate is limited by how quickly the brain can process or produce information. For example, one study found that people can generally understand audio recordings that are sped up to 120%. 

While the rate of speech is relatively constant, the information density and speaking rate can vary. For example, the information density of Basque is 5 bits per syllable, while Vietnamese is 8 bits per syllable. 

Current state of the art fibre optic cables can transmit up to 10 terabits a second. 

That's probably a wild overestimate for AI communication though. More relevant bottlenecks are limits on processing informations [plausibly more in the megabits range], limits on transferability of activation vectors (but training could improve this). 

Replies from: carl-feynman
comment by Carl Feynman (carl-feynman) · 2024-11-06T17:23:05.197Z · LW(p) · GW(p)

A fascinating recent paper on the topic of human bandwidth  is https://arxiv.org/abs/2408.10234.  Title and abstract:

The Unbearable Slowness of Being

Jieyu Zheng, Markus Meister

This article is about the neural conundrum behind the slowness of human behavior. The information throughput of a human being is about 10 bits/s. In comparison, our sensory systems gather data at an enormous rate, no less than 1 gigabits/s. The stark contrast between these numbers remains unexplained. Resolving this paradox should teach us something fundamental about brain function: What neural substrate sets this low speed limit on the pace of our existence? Why does the brain need billions of neurons to deal with 10 bits/s? Why can we only think about one thing at a time? We consider plausible explanations for the conundrum and propose new research directions to address the paradox between fast neurons and slow behavior.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-09-29T13:15:43.912Z · LW(p) · GW(p)

God exists because the most reasonable take is the Solomonoff Prior. 

A funny consequence of that is that Intelligent Design will have a fairly large weight in the Solomonoff prior. Indeed the simulation argument can be seen as a version of Intelligent Design. 

The Abrahamic God hypothesis is still substantially downweighted because it seems to involve many contigent bits - i.e noisy random bits that can't be compressed. The Solomonoff prior therefore has to downweight them. 

Replies from: Mitchell_Porter, Nate Showell
comment by Mitchell_Porter · 2024-09-29T16:47:28.119Z · LW(p) · GW(p)

Please demonstrate that the Solomonoff prior favors simulation.

Replies from: thomas-kwa, alexander-gietelink-oldenziel
comment by Thomas Kwa (thomas-kwa) · 2024-09-29T20:32:02.187Z · LW(p) · GW(p)

See e.g. Xu (2020) [LW · GW] and recent criticism [LW · GW].

Replies from: Mitchell_Porter
comment by Mitchell_Porter · 2024-09-30T00:48:39.816Z · LW(p) · GW(p)

I was expecting an argument like "most of the probability measure for a given program, is found in certain embeddings of that program in larger programs". Has anyone bothered to make a quantitative argument, a theorem, or a rigorous conjecture which encapsulates this claim?

Replies from: thomas-kwa
comment by Thomas Kwa (thomas-kwa) · 2024-09-30T04:13:59.637Z · LW(p) · GW(p)

I don't think that statement is true since measure drops off exponentially with program length.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-09-29T17:12:33.129Z · LW(p) · GW(p)

This is a common belief around here. Any reason you are skeptical?

Replies from: Mitchell_Porter
comment by Mitchell_Porter · 2024-09-30T04:50:28.952Z · LW(p) · GW(p)

Thomas Kwa just provided a good reason: "measure drops off exponentially with program length". So embeddings of programs within other programs - which seems to be what a simulation is, in the Solomonoff framework - are considered exponentially unlikely. 

edit: One could counterargue that programs simulating programs increase exponentially in number. Either way, I want to see actual arguments or calculations.  

Replies from: thomas-kwa
comment by Thomas Kwa (thomas-kwa) · 2024-09-30T22:36:47.443Z · LW(p) · GW(p)

I just realized what you meant by embedding-- not a shorter program within a longer program, but a short program that simulates a potentially longer (in description length) program.

As applied to the simulation hypothesis, the idea is that if we use the Solomonoff prior for our beliefs about base reality, it's more likely to be laws of physics for a simple universe containing beings that simulate this one as it is to be our physics directly, unless we observe our laws of physics to be super simple. So we are more likely to be simulated by beings inside e.g. Conway's Game of Life than to be living in base reality.

I think the assumptions required to favor simulation are something like

  • there are universes with physics 20 bits (or whatever number) simpler than ours in which intelligent beings control a decent fraction >~1/million of the matter/space
  • They decide to simulate us with >~1/million of their matter/space
    • There has to be some reason the complicated bits of our physics are more compressible by intelligences than by any compression algorithms simpler than their physics; they can't just be iterating over all permutations of simple universes in order to get our physics
    • But this seems fairly plausible given that constructing laws of physics is a complex problem that seems way easier if you are intelligent. 

Overall I'm not sure which way the argument goes. If our universe seems easy to efficiently simulate and we believe the Solomonoff prior, this would be huge evidence for simulation, but maybe we're choosing the wrong prior in the first place and should instead choose something that takes into account runtime.

comment by Nate Showell · 2024-09-29T20:19:51.095Z · LW(p) · GW(p)

Why are you a realist about the Solomonoff prior instead of treating it as a purely theoretical construct?

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-04-10T10:24:10.297Z · LW(p) · GW(p)

Clem's Synthetic- Physicalist Hypothesis

The mathematico-physicalist hypothesis states that our physical universe is actually a piece of math. It was famously popularized by Max Tegmark. 

It's one of those big-brain ideas that sound profound when you first hear about it, then you think about it some more and you realize it's vacuous. 

Recently, in a conversation with Clem von Stengel they suggested a version of the mathematico-physicalist hypothesis that I find provoking. 

Synthetic mathematics 

'Synthetic' mathematics is a bit of weird name. Synthetic here is opposed to 'analytic' mathematics, which isn't very meaningful either. It has nothing to do with the mathematical field of analysis. I think it's supposed to a reference to Kant's synthetic/  apriori/ a posteriori. The name is probably due to Lawvere. 

nLab:

"In “synthetic” approaches to the formulation of theories in mathematics the emphasis is on axioms that directly capture the core aspects of the intended structures, in contrast to more traditional “analytic” approaches where axioms are used to encode some basic substrate out of which everything else is then built analytically."

If you read synthetic read 'Euclidean'. As in - Euclidean geometry is a bit of an oddball field of mathematics, despite being the oldest - it defines points and lines operationally instead of out of smaller pieces (sets). 

In synthetic mathematics you do the same but for all the other fields of mathematics. We have synthetic homotopy theory (aka homotopy type theory), synthetic algebraic geometry, synthetic differential geometry, synthetic topology etc. 

A type in homotopy type theory is solely defined by its introduction rules and elimination rules (+ univalence axiom). It means a concept it defined solely by how it is used - i.e. operationally. 

Agent-first ontology & Embedded Agency

Received opinion is that Science! says there is nothing but Atoms in the Void. Thinking in terms of agents, first-person view concepts like I and You, actions & observations, possibilities & interventions is at best an misleading approximation at worst a degerenerate devolution to cavemen thought. The surest sign of a kook is their insistence that quantum mechanics proves the universe is conscious. 

But perhaps the way forward is to channel our inner kook. What we directly observe is qualia, phenomena, actions not atoms in the void. The fundamental concept is not atoms in the void, but agents embedded in environments

(see also Cartesian Frames, Infra-Bayesian Physicalism & bridge rules, UDASSA)

Physicalism 

What would it look like for our physical universe to be a piece of math? 

Well internally to synthetic mathematical type theory there would be something real - the universe is a certain type. A type such that it 'behaves' like a 4-dimensional manifold (or something more exotic like 1+1+3+6 rolled up Calabi-Yau monstrosities). 

The type is defined by introduction and elimination rules - in other words operationally: the universe is what one can *  do  *with it . 

Actually instead of thinking of the universe as a fixed static object we should be thinking of an embedded agent in a environment-universe. 

That is we should be thinking of an  * interface *

[cue: holographic principle]

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-03-26T11:59:06.995Z · LW(p) · GW(p)

Know your scientific competitors. 

In trading, entering a market dominated by insiders without proper research is a sure-fire way to lose a lot of money and time.  Fintech companies go to great lengths to uncover their competitors' strategies while safeguarding their own.

A friend who worked in trading told me that traders would share subtly incorrect advice on trading Discords to mislead competitors and protect their strategies.

Surprisingly, in many scientific disciplines researchers are often curiously incurious about their peers' work.

The long feedback loop for measuring impact in science, compared to the immediate feedback in trading, means that it is often strategically advantageous to be unaware of what others are doing. As long as nobody notices during peer review it may never hurt your career.  

But of course this can lead people to do completely superflueous, irrelevant  & misguided work. This happens often. 

Ignoring competitors in trading results in immediate financial losses. In science, entire subfields may persist for decades, using outdated methodologies or pursuing misguided research because they overlook crucial considerations. 

Replies from: Viliam
comment by Viliam · 2024-03-26T12:47:17.296Z · LW(p) · GW(p)

Makes sense, but wouldn't this also result in even fewer replications (as a side effect of doing less superfluous work)?

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-04-14T13:16:42.280Z · LW(p) · GW(p)

Four levels of information theory

There are four levels of information theory. 

Level 1:  Number Entropy 

Information is measured by Shannon entropy

Level 2: Random variable 

look at the underlying random variable ('surprisal')   of which entropy is the expectation.

Level 3: Coding functions

Shannon's source coding theorem says entropy of a source  is the expected number of bits for an optimal encoding of samples of .

Related quantity like mutual information, relative entropy, cross entropy, etc can also be given coding interpretations. 

Level 4: Epsilon machine (transducer)

On level 3 we saw that entropy/information actually reflects various forms of (constrained) optimal coding. It talks about the codes but it does not talk about how these codes are  implemented. 

This is the level of Epsilon machines, more precisely epsilon transducers. It says not just what the coding function is but how it is (optimally) implemented mechanically. 

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-02-05T18:04:48.615Z · LW(p) · GW(p)

Idle thoughts about UDASSA I: the Simulation hypothesis 

I was talking to my neighbor about UDASSA the other day. He mentioned a book I keep getting recommended but never read where characters get simulated and then the simulating machine is progressively slowed down. 

One would expect one wouldn't be able to notice from inside the simulation that the simulating machine is being slowed down.

This presents a conundrum for simulation style hypotheses: if the simulation can be slowed down 100x without the insiders noticing, why not 1000x or 10^100x or quadrilliongoogolgrahamsnumberx? 

If so - it would mean there is a possibly unbounded number of simulations that can be run. 

Not so, says UDASSA. The simulating universe is also subject to UDASSA. This imposes a restraint on the size and time period that the simulating universe is in. Additionally, ultraslow computation is in conflict with thermodynamic decay - fighting thermodynamic decay costs descriptiong length bits which is punished by UDASSA. 

I conclude that this objection to simulation hypotheses are probably answered by UDASSA. 

Idle thoughts about UDASSA II: Is Uploading Death?

There is an argument that uploading doesn't work since encoding your brain into a machine incurs a minimum amount of encoding bits. Each bit is a 2x less Subjective Reality Fluid according to UDASSA [LW · GW] so even a small encoding cost would mean certain subjective annihiliation. 

There is something that confuses me in this argument. Could it not be possible to encode one's subjective experiences even more efficiently than in a biological body? This would make you exist MORE in an upload.

OTOH it becomes a little funky again when there are many copies as this increases the individual coding cost (but also there are more of you sooo). 

Replies from: Dagon
comment by Dagon · 2024-02-09T20:32:12.179Z · LW(p) · GW(p)

In most conceptions of simulation, there is no meaning to "slowed down", from the perspective of the simulated universe.  Time is a local phenomenon in this view - it's just a compression mechanism so the simulators don't have to store ALL the states of the simulation, just the current state and the rules to progress it.   

Note that this COULD be said of a non-simulated universe as well - past and future states are determined but not accessible, and the universe is self-discovering them by operating on the current state via physics rules.  So there's still no inside-observable difference between simulated and non-simulated universes.

UDASSA seems like anthropic reasoning to include Boltzmann Brain like conceptions of experience.  I don't put a lot of weight on it, because all anthropic reasoning requires an outside-view of possible observations to be meaningful.

And of course, none of this relates to upload, where a given sequence of experiences can span levels of simulation.  There may or may not be a way to do it, but it'd be a copy, not a continuation.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-02-09T22:47:43.010Z · LW(p) · GW(p)

The point you make in the your first paragraph is contained in the original shortform post. The point of the post is exactly that an UDASSA-style argument can nevertheless recover something like a 'distribution of likely slowdown factors'. This seems quite curious.

I suggest reading Falkovich's post on UDASSA to get a sense whats so intriguing abouy the UDASSA franework.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-09-10T22:30:30.941Z · LW(p) · GW(p)

The Virtue of Comparison Shopping 
Comparison shopping, informed in-depth reviewing, answering customer surveys plausibly have substantial positive externalities. It provides incentives through local actors, avoids preference falsification or social desirability bias, and is non-coercive & market-based. 

Plausibly it is even has a better social impact than many kinds of charitable donations or direct work. This is not so hard since it seems that the latter contains many kinds of interventions that have neglibible or even negative impact. 

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-08-15T14:43:15.693Z · LW(p) · GW(p)

Gaussian Tails and Exceptional Performers

West African athletes dominate sprinting events, East Africans excel in endurance running, and despite their tiny population Icelanders have shown remarkable prowess in weightlifting competitions.  We examine the Gaussian approximation for a simple additive genetic model for these observations. 

The Simple Additive Genetic Model

Let's begin by considering a simple additive genetic model. In this model, a trait T is influenced by n independent genes, each contributing a small effect, along with environmental factors. We can represent this mathematically as:

T = G₁ + G₂ + ... + Gₙ + E

Where Gᵢ represents the effect of the i-th gene, and E represents environmental factors.

The Central Limit Theorem (CLT) suggests that the sum of many independent random variables, each with finite mean and variance, will approach a normal (Gaussian) distribution, regardless of the underlying distribution of the individual variables. Mathematically, if we have n independent random variables X₁, X₂, ..., Xₙ, each with mean μᵢ and variance σᵢ², then their sum S = X₁ + X₂ + ... + Xₙ will approach a normal distribution as n increases:

(S - μ) / (σ√n) → N(0,1) as n → ∞

Where μ = Σμᵢ and σ² = Σσᵢ²

This model seems particularly applicable to sports like running and weightlifting, which are widely practiced around the world and rely on fundamental physiological traits. The global nature of these sports suggests that differences in performance are less likely to be solely due to cultural or environmental factors.

Caution (can be skipped)


However, we must exercise caution in interpreting these models. While genetic factors likely play a role in patterns of exceptional performance, we must be wary of inferring purely genetic origins. Environmental and cultural factors can have significant impacts. For example:

  1. The success of Ethiopian long-distance runners may be partly attributed to high-altitude training in the Ethiopian highlands, a practice that has been adopted with success by athletes from other regions.
  2. Cultural emphasis on certain sports in specific regions can lead to more robust talent identification and training programs.
  3. Socioeconomic factors can influence access to training facilities, nutrition, and coaching.

Moreover, the simple additive genetic model assumes that traits are purely additively genetic and normally distributed, which may not always be the case, especially at the extremes. Gene-environment interactions and epistasis (gene-gene interactions) may become more significant at these extremes, leading to deviations from the expected Gaussian distribution.

Minority Overrepresentation in Extreme Performance

The phenomenon of minority overrepresentation in certain fields of extreme performance provides intriguing insights into the nature of trait distributions. A striking example of this is the dominance of East African runners in marathon events. More specifically, it is East Africans that usually win major marathons. Since the 1968 Olympics, men and women from Kenya and Ethiopia have dominated the 26.2-mile event. Since 1991, the men's winner at the Boston Marathon has been either a Kenyan or Ethiopian 26 of the last 29 times. East African women have worn the laurel wreath 21 times in the last 24 years at Boston. Upsets do happen, when in 2018 a crazy Japanese amateur won the Men's marathon, and an American woman won the Woman's marathon. 

To understand why even small differences in mean genetic potential can lead to substantial overrepresentation at the extreme tails of performance, let's examine the mathematics of Gaussian distributions.

Consider two populations with normally distributed traits, G and H, where H has a higher mean (m_H) than G (m_G), but they share the same standard deviation s. We can calculate the ratio of probabilities at a given level of standard deviation using the formula:

R = e^(kd - d²/2)

Where:

  • k is the number of standard deviations from the mean of G
  • d = (m_H - m_G) / s, the difference between means in units of standard deviation

Let's consider four cases: d = 2/3, d = 4/5, d = 1, and d = 5/4. Here's how the overrepresentation ratio R changes at different standard deviations:

kd = 2/3d = 4/5d = 1d = 5/4
11.39561.49181.64871.9477
22.71823.32014.48176.0496
35.29337.389112.182518.7874
410.308416.444633.115558.3442
520.085536.598290.0171181.2721
639.109581.4509244.6919562.9412

To put these numbers in perspective, let's consider the rarity of individuals at each standard deviation:

kPercentage1 in X
115.87%6.3
22.28%44
30.13%769
40.003%33,333
50.00003%3,333,333
60.0000001%1,000,000,000

In other words, a '6 Sigma' event is a 1 in a billion event. A very naive inference would predict that there are 8 people in the world at this level. [1]

Now, let's consider the case of East African marathon runners. The population of East Africa is approximately 250 million, while the global population is about 8 billion. This means East Africans represent about 3.125% of the world population.

For world-record level running performance, we're likely looking at something around k=5 or k=6. At these levels:

  • For d = 2/3: East Africans would be overrepresented by a factor of 20-39
  • For d = 4/5: East Africans would be overrepresented by a factor of 37-81
  • For d = 1: East Africans would be overrepresented by a factor of 90-245
  • For d = 5/4: East Africans would be overrepresented by a factor of 181-563

Given that East Africans represent about 3.125% of the world population, if they were winning close to 100% of major marathons, this would suggest a d value between 1 and 5/4. However, as we'll discuss in the next section, we should be cautious about drawing definitive conclusions from these mathematical models alone.

The "Tails Come Apart" Phenomenon

(see also Tails Come Apart)

While the Gaussian model explains many observations, extreme outliers often deviate from this model. 

  1. Different Causes: Extreme outliers may result from fundamentally different mechanisms than those governing the main distribution. For example, the tallest man in recorded history, Robert Wadlow, reached an extraordinary height of 8'11" (2.72m) due to a rare condition causing excessive growth hormone production.
  2. Breakdown of the Additive Model: At extreme values, the simple additive genetic model may break down. Interactions between genes (epistasis) or between genes and environment may become more significant, leading to deviations from the expected Gaussian distribution.
  3. Gaussian approximation fails at the tails 
    As an example, for a sum of Bernouli trials the CLT approximation overestimates the fatness of the tails compared to an exact calculation or large deviation theory, see here for a GPT calculation. 

Final Thoughts

We posited a simple additive genetic model for long-distance running, used the Central Limit Theorem approximation to estimate likelihood for extreme outliers for different mean populations. Using observed frequencies of extreme outliers and making the possibly questionable assumption that the long tails are in fact well-approximate by the Bell curve we obtained an estimate for the difference in mean traits. 

Note that if you knew the ratio over extreme performers and the difference in means for two populations you can use that to test if the distribution is in fact well appproximated by a Bell curve at the tails. 

More interestingly, if one is measuring a proxy trait and wonders whether to what degree this explains performance, observing a higher mean subpopulation overrepresented at tails can give an indication to what degree this trait is relevant for predicting extreme performance. We leave further inferences to the reader. 

 

  1. ^

    Surprisingly to me, this naive extrapolation is just about compatible with some seemingly outrageous claims of high IQs. Marilyn vos Savant at 188 IQ (Guinness World Record) and 195 IQ for Christopher Langan. This is ~ within 6 standard deviations for a Bell curve of mean a 100 and standard deviation 15. One would think these are meaningless numbers, more a result of test ceilings, test error, test inaccuracy and simple lies but perhaps there is something to it. 

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-08-07T12:39:47.392Z · LW(p) · GW(p)

tl;dr 

Salmon swimming upstream to their birthing grounds to breed may be that rare form of group selection. 

Pure Aryan Salmon

Salmon engage in anodromous reproduction; they risk their lives to swim up rivers to return to their original place of birth and reproduce there. 

Most species of salmon die there, only reproducing at the birthing grounds. Many don't make it at all. The ones that survive the run upstream will die shortly after, a biologically triggered death sentence. If the cost is immense - the benefits must be even greater.

The more upstream the saver. The solid magikarp babies are saver the more upstream as the lakes and pools and fountainheads high up support less predators. Downstream there are more nutrients so as the magikarps grow larger they move downstream. This makes perfect sense and plausibly explain the behaviour. but it recently struck me that there is also a group selection story here. 

Not all species are created equal. Some species are simply more biologically fit than others. For instance, the native Australian fauna generically canot compete with Old World fauna. Ditto for many island fauna. 

How could this be? Aren't all animals equally selected to reproduce? Not all environments are the same. Maintaining biological vigor in the face of mutational load is difficult. The amount of purifying selection is a major factor in the level of mutational load and biological vigor of a species. [1]

Group selection is a hypothesized form of selection that acts at the level of the group rather than that of the individual. Group selection has gotten a bad rap. In the vast majority of the cases hypothesized group selection is better explained by selection at the level of the gene. The math of group selection shakes out in such a way that it is rarely a strong enough effect. 

An exception that is sometimes mentioned is the emergence and exctinction of asexual species. Although being asexual makes one more fecund in the short run, in the long run it prevents one from doing sexual recombination.  Asexual species universally seem to have come into being very recently. They likely go extinct due to lack of genetic diversity and attendant mutational load catastrophe and/or losing arms races with parasites. 

The equilibrium of swimming back to the ancestral pool to reproduct would seem like an insane practice on the face of it. Why go through all this effort?

One explanationc could be that anodromous reproduction is a stable game-theoretic equilibrium in which the selective pressure on the salmon species is higher encouraging higher biological fitness. 

Salmon that deviate from the practice either don't get to mate or if they mate they are reproductively isolated from the other salmon and become a different species. 

This group selection hypothesis makes predictions about the biological vigor and mutational load of salmon and could be, is likely, false. 

 

  1. ^

    Another important factor for the strength of selection is the effective population size which I will pass over in silence for the moment. But if you're interested in this topic I suggest you acquaint yourself with this topic as it is probably the most important and least understood concept in evolutionary biology. 

Replies from: Raemon, Richard_Kennaway, mateusz-baginski
comment by Raemon · 2024-08-07T17:55:42.200Z · LW(p) · GW(p)

I am confused about what I'm reading. The magikarp gave me a doubletake and like "wait, are magikarp also just a totally real fish?" but after some googling it seems like "nope, that's really just a pokemon", and now I can't tell if the rest of the post is like a parody or what.

Replies from: habryka4
comment by habryka (habryka4) · 2024-08-07T19:13:00.480Z · LW(p) · GW(p)

This post feels like raw GPT-3 output. It's not even GPT-3.5 or GPT-4 level, which makes this additionally confusing. 

Maybe a result of playing around with base models?

Replies from: D0TheMath
comment by Garrett Baker (D0TheMath) · 2024-08-07T19:45:10.347Z · LW(p) · GW(p)

This seems fairly normal for an Alexander post to me (actually, more understandable than the median Alexander shortform). I think the magikarp is meant to be 1) an obfuscation of salamon, and 2) a reference to solid gold magikarp.

@Raemon [LW · GW

Replies from: habryka4
comment by habryka (habryka4) · 2024-08-07T20:53:22.949Z · LW(p) · GW(p)

After rereading it like 4 times I am now less convinced it's GPT output. I still feel confused about a lot of sentences, but I think half of it was just the lack of commas in sentences like "One explanationc could be that anodromous reproduction is a stable game-theoretic equilibrium in which the selective pressure on the salmon species is higher encouraging higher biological fitness".

comment by Richard_Kennaway · 2024-08-07T16:01:41.789Z · LW(p) · GW(p)

The solid magikarp babies are saver the more upstream as the lakes and pools and fountainheads high up support less predators. Downstream there are more nutrients so as the magikarps grow larger they move downstream.

Please tell us more about the magikarps.

comment by Mateusz Bagiński (mateusz-baginski) · 2024-08-07T13:09:13.117Z · LW(p) · GW(p)

Asexual species universally seem to have come into being very recently. They likely go extinct due to lack of genetic diversity and attendant mutational load catastrophe and/or losing arms races with parasites.

Bdelloidea are an interesting counterexample: they evolved obligate parthenogenesis ~25 mya.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-08-07T13:39:19.507Z · LW(p) · GW(p)

My understanding from reading Mitochondria: Power, Sex, Suicide is that they are not truly asexual but turn out to do some sexual recombination. I don't remember the details and I'm not an expert though so wouldn't put my hand in the fire for it.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-06-13T17:19:13.259Z · LW(p) · GW(p)

Why do people like big houses in the countryside /suburbs?

Empirically people move out to the suburbs/countryside when they get children and/or gain wealth. Having a big house with a large yard is the quintessential American dream. 

but why? Dense cities are economoically more productive, commuting is measurably one of the worst factors for happiness and productivity. Raising kids in small houses is totally possible and people have done so at far higher densities in the past. 

Yet people will spend vast amounts of money on living in a large house with lots of space - even if they rarely use most rooms. Having a big house is almost synonymous with wealth and status. 

Part of the reason may be an evolved disease response. In the past, the most common way to die was as a child dieing to a crowd-disease. There was no medicine that actually worked yet wealthier people had much longer lifespans and out reproduced the poor (see Gregory Clark). The best way to buy health was to move out of the city (which were population sinks until late modernity) and live in a large aired house. 

It seems like an appealing model. On the other hand, there are some obvious predicted regularities that aren't observed to my knowledge. 

Replies from: adam-shai, Dagon
comment by Adam Shai (adam-shai) · 2024-06-13T20:01:35.535Z · LW(p) · GW(p)

I can report my own feelings with regards to this. I find cities (at least the American cities I have experience with) to be spiritually fatiguing. The constant sounds, the lack of anything natural, the smells - they all contribute to a lack of mental openness and quiet inside of myself.

The older I get the more I feel this.

Jefferson had a quote that might be related, though to be honest I'm not exactly sure what he was getting at:
 

I think our governments will remain virtuous for many centuries; as long as they are chiefly agricultural; and this will be as long as there shall be vacant lands in any part of America. When they get piled upon one another in large cities, as in Europe, they will become corrupt as in Europe. Above all things I hope the education of the common people will be attended to; convinced that on their good sense we may rely with the most security for the preservation of a due degree of liberty.

One interpretation of this is that Jefferson thought there was something spiritually corrupting of cities. This supported by another quote:
 


I view great cities as pestilential to the morals, the health and the liberties of man. true, they nourish some of the elegant arts; but the useful ones can thrive elsewhere, and less perfection in the others with more health virtue & freedom would be my choice.

 

although like you mention, there does seem to be some plausible connection to disease.

comment by Dagon · 2024-06-13T20:47:14.493Z · LW(p) · GW(p)

Note that it could easily be culturally evolved, not genetically.  I think there's a lot of explanatory power in the land=status cultural belief as well.  But really, I think there's a typical mind fallacy that blinds you to the fact that many people legitimately and truly prefer those tradeoffs over denser city living.  Personally, my tastes (and the character of many cities' cores) have noticeably changed over my lifetime - in my youth, I loved the vibrance and variety, and the relatively short commute of being in a city.  Now, I value the privacy and quiet that suburban living (still technically in-city, but in a quiet area) gets me.

More importantly, for many coastal American cities, it's simply not true that people pay a lot to live in the suburbs.  Even in the inflationary eras of the 1980s, a standalone single-family house in an area where most neighbors are rich and value education is more investment than expense (or was when they bought the house.  Who knows whether it will be in the future).

I don't have good answers for the commuting sucks and density correlates with productivity arguments, except that revealed preference seems to contradict those as being the most important things.  Also, the measurements I've seen seem to include a range of circumstances that make it hard to separate the actual motivations.  Living by choice in "the nice" suburbs is likely a very different experience with different desirability than living in a cheap apartment with a long commute because you can't afford to live in the city.  I'd be interested to see same-age, same-family-situation, similar wealth comparisons of city and suburb dwellers.   

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-04-14T18:37:37.209Z · LW(p) · GW(p)

Why (talk-)Therapy 

Therapy is a curious practice.  Therapy sounds like a scam, quackery, pseudo-science but it seems RCT consistently show therapy has benefits above and beyond medication & placebo. 

Therapy has a long history. The Dodo verdict states that it doesn't matter which form of therapy you do - they all work equally well. It follows that priests and shamans served the functions of a therapist.  In the past, one would confessed ones sins to a priest, or spoken with the local shaman. 

There is also the thing that therapy is strongly gendered (although this is changing), both therapists and their clientele lean female. 

Self-Deception 

Many forecasters will have noticed that their calibration score tanks the moment they try to predict salient facts about themselves. We are not-well calibrated about our own beliefs and desires. 

Self-Deception is very common, arguably inherent to the human condition. There are of course many Hansonian reasons for this. I refer the reader to the Elephant and the Brain. Another good source would be Robert Trivers. These are social reasons for self-deception. 

It is also not implausible that there are non-social reasons for self-deception. Predicting one-self perfectly can in theory lead one to get stuck in Procrastination Paradoxes. Whether this matters in practice is unclear to me but possible. Exuberant overconfidence is another case that seems like a case of self-deception. 

Self-deception can be very useful, but one still pays the price for being inaccurate. The main function of talk-therapy seems to be to have a safe, private space in which humans can temporarily step out of their self-deception and reasses more soberly where they are at. 

It explains many salient features of talk- therapy: the importance of talking extensively to another person that is (professionally) sworn to secrecy and therefore unable to do anything with your information. 

Replies from: lcmgcd
comment by lukehmiles (lcmgcd) · 2024-04-14T18:54:17.762Z · LW(p) · GW(p)

I suspect that past therapists existed in your community and knew what you're actually like so were better able to give you actual true information instead of having to digest only your bullshit and search for truth nuggets in it.

Furthermore, I suspect they didn't lose their bread when they solve your problem! We have a major incentive issue in the current arrangement!

Replies from: M. Y. Zuo, alexander-gietelink-oldenziel
comment by M. Y. Zuo · 2024-04-14T19:13:14.965Z · LW(p) · GW(p)

There's a market for lemons problem, similar to the used car market, where neither the therapist nor customer can detect all hidden problems, pitfalls, etc., ahead of time. And once you do spend enough time to actually form a reasonable estimate there's no takebacks possible.

So all the actually quality therapists will have no availability and all the lower quality therapists will almost by definition be associated with those with availability.

Edit: Game Theory suggests that you should never engage in therapy or at least never with someone with available time, at least until someone invents the certified pre-owned market.

Replies from: ChristianKl, D0TheMath, alexander-gietelink-oldenziel
comment by ChristianKl · 2024-04-16T11:46:39.601Z · LW(p) · GW(p)

Edit: Game Theory suggests that you should never engage in therapy or at least never with someone with available time, at least until someone invents the certified pre-owned market.

That would be prediction-based medicine [LW · GW]. It works in theory, it's just that someone would need to put it into practice. 

comment by Garrett Baker (D0TheMath) · 2024-04-14T19:25:09.995Z · LW(p) · GW(p)

This style of argument proves too much. Why not see this dynamic with all jobs and products ever?

Replies from: lcmgcd
comment by lukehmiles (lcmgcd) · 2024-04-14T19:27:20.570Z · LW(p) · GW(p)

Have you ever tried hiring someone or getting a job? Mostly lemons all around (apologies for the offense, jobseekers, i'm sure you're not the lemon)

Replies from: localdeity
comment by localdeity · 2024-04-15T00:25:55.291Z · LW(p) · GW(p)

Yup.  Many programmer applicants famously couldn't solve FizzBuzz.  Which is probably because:

[skipping several caveats and simplifying assumptions]

Now, when you get those 200 resumes, and hire the best person from the top 200, does that mean you’re hiring the top 0.5%?

“Maybe.”

No. You’re not. Think about what happens to the other 199 that you didn’t hire.

They go look for another job.

That means, in this horribly simplified universe, that the entire world could consist of 1,000,000 programmers, of whom the worst 199 keep applying for every job and never getting them, but the best 999,801 always get jobs as soon as they apply for one. So every time a job is listed the 199 losers apply, as usual, and one guy from the pool of 999,801 applies, and he gets the job, of course, because he’s the best, and now, in this contrived example, every employer thinks they’re getting the top 0.5% when they’re actually getting the top 99.9801%.

Replies from: D0TheMath
comment by Garrett Baker (D0TheMath) · 2024-04-15T01:20:13.693Z · LW(p) · GW(p)

But such people are very obvious. You just give them a FizzBuzz test! This is why we have interviews, and work-trials.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-04-14T19:17:06.606Z · LW(p) · GW(p)

If therapist quality would actually matter why don't we see this reflected in RCTs?

Replies from: ChristianKl
comment by ChristianKl · 2024-04-16T11:49:03.464Z · LW(p) · GW(p)

We see it reflected in RCTs. One aspect of therapist quality is for example therapist empathy and empathy is a predictor for treatment outcomes

The style of therapy does not seem to be important according to RCTs but that doesn't mean that therapist skill is irrelevant. 

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-04-16T15:54:17.223Z · LW(p) · GW(p)

Thank you practicing the rationalist virtue of scholarship Christian. I was not aware of this paper. 

You will have to excuse me for practicing rationalist vice and not believing nor investigating further this paper. I have been so traumatized by the repeated failures of non-hard science, I reject most social science papers as causally confounded p-hacked noise unless it already confirms my priors or is branded correct by somebody I trust. 

Replies from: ChristianKl
comment by ChristianKl · 2024-04-16T20:04:00.795Z · LW(p) · GW(p)

As far as this particular paper goes I just searched for one on the point in Google Scholar. 

I'm not sure what you believe about Spencer Greenberg but he has two interviews with people who believe that therapist skills (where empathy is one of the academic findings) matter:

https://podcast.clearerthinking.org/episode/070/scott-miller-why-does-psychotherapy-work-when-it-works-at-all/

https://podcast.clearerthinking.org/episode/192/david-burns-cognitive-behavioral-therapy-and-beyond/

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-04-14T19:08:15.778Z · LW(p) · GW(p)

I internalized the Dodo verdict and concluded that the specific therapist or therapist style didn't matter anyway. A therapist is just a human mirror. The answer was inside of you all along Miles

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-02-09T01:57:46.104Z · LW(p) · GW(p)

[This is joint thinking with Sam Eisenstat. Also thanks to Caspar Oesterheld for his thoughtful comments. Thanks to Steve Byrnes for pushing me to write this out.]


The Hyena problem in long-term planning  

Logical induction is a nice framework to think about bounded reasoning. Very soon after the discovery of logical induction people tried to make logical inductor decision makers work. This is difficult to make work [LW · GW]: one of two obstacles is

Obstacle 1: Untaken Actions are not Observable

Caspar Oesterheld brilliantly solved this problem by using auction markets in defining his bounded rational inductive agents. 

The BRIA framework is only defined for single-step/  length 1 horizon decisions. 

What about the much more difficult question of long-term planning? I'm going to assume you are familiar with the BRIA framework. 

 

Setup: we have a series of decisions D_i, and rewards R_i, i=1,2,3... where rewards R_i can depend on arbitrary past decisions. 

We again think of an auction market M of individual decisionmakers/ bidders.

There are a couple design choices to make here:

  • bidders directly bet for an action A in a decision D_i or bettors bet for rewards on certain days.
  • total observability or partial observability. 
  • bidders can bid conditional on observations/ past actions or not
  • when can the auction be held? i.e. when is an action/ reward signal definitely sold?

To do good long-term planning it should be possible for one of the bidders or a group of bidders to commit to a long-term plan, i.e. a sequence of actions. They don't want to be outbid in the middle of their plan. 

There are some problems with the auction framework: if bids for actions can't be combined then an outside bidder can screw up the whole plan by making a slighly higher bid for an essential part of the plan. This look like ADHD. 

How do we solve this? One way is to allow a bidder or group of bidders to bid for a whole sequence of actions for a single lumpsum. 

  • One issue is that we also have to determine how the reward gets awarded. For instance the reward could be very delayed. This could be solved by allowing for bidding for a reward signal R_i on a certain day conditional on a series of actions. 

There is now an important design choice left. When a bidder  owns a series of actions A=a_1,..,a_k (some of the actions in the future, some already in the past) when there is another bid  from another bidder  on future actions 

  • is bidder  forced to sell their contract on  to  if the bid is high enough ? [higher than the original bid]

Both versions seem problematic:

  • if they don't have to there is an Incumbency Advantage problem. An initially rich bidder can underbid for very long horizons and use the steady trickle of cash to prevent any other bidders from ever being to underbid any actions. 
  • Otherwise there is the Hyena problem. 

The Hyena Problem 

Imagine the following situation: on Day 1 the decisionmaker has a choice of actions. The highest expected value action is action a. If action a is made on Day 2 a fair coin is flipped. On Day 3 the reward is paid out. 

If the coin was heads, 15 reward is paid out.

If the coin was tails, 5 reward is paid out.

The expected value is therefore 10. This is higher (by assumption) than the other unnamed actions. 

However if the decisionmaker is a long-horizon BRIA with forced sales there is a pathology. 

A sensible bidder is willing to pay up to 10 utilons for the contracts on the day 3 reward conditional on action a. 

However, with a forced sale mechanism on Day 2 a 'Hyena bidder'  can come that will 'attempt to steal the prey'. 

The Hyena bidder bids >10 for the contract if the coin comes up heads on Day 2 but doesn't bid anything for the contract if the coin comes up tails.

This is a problem since the expected value of the action a for the sensible bidder goes down, so the sensible bidder might no longer bid for the action that maximizes expected value for the BRIA. The Hyena bidder screws up the credit allocation. 

some thoughts:

  • if the sensible bidder is able to make bids conditional on the outcome of the coin flip that prevents Hyena bidder. This is a bit weird though because it would mean that the sensible bidder must carry around lots of extraneous non-necessary information instead of just caring about expected value.
  • perhaps this can alleviated by having some sort of 'neo-cortex' separate logical induction markets that is incentivized to have accurate beliefs. This is difficult to get right: the prediction market needs to be incentivized to get accurate on beliefs that are actually action relevant, not random beliefs - if the prediction market and the auction market are connected too tightly you might run the risk of getting into the old problems of Logical Inductor Decision makers. [they underexplore since untaken action are not observed]. 
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-06-30T04:35:09.150Z · LW(p) · GW(p)

Latent abstractions Bootlegged.

Let  be random variables distributed according to a probability distribution  on a sample space 

Defn. A (weak) natural latent of  is a random variable  such that

(i)   are independent conditional on 

(ii) [reconstructability]   for all 

[This is not really reconstructability, more like a stability property. The information is contained in many parts of the system... I might also have written this down wrong]

Defn. A strong natural latent  additionally satisfies 

Defn. A natural latent is noiseless if ?

 ??

[Intuitively,  should contain no independent noise not accoutned for by the ]

Causal states

Consider the equivalence relation on tuples  given   if for all  

We call the set of equivalence relation   the set of causal states.

By pushing forward the distribution  on  along the quotient map 

This gives a noiseless (strong?) natural latent .

Remark. Note that Wentworth's natural latents are generalizations of Crutchfield causal states (and epsilon machines).

Minimality and maximality 

Let  be random variables as before and let  be a weak latent. 

Minimality Theorem for Natural Latents.  Given any other variable  such that  the   are independent conditional on  we have the following DAG

 

i.e.   

[OR IS IT for all  ?]

Maximality Theorem for Natural Latents.  Given any other variable  such that the reconstrutability property holds with regard to  we have 

 

Some other things:

  • Weak latents are defined up to isomorphism? 
  • noiseless weak (strong?) latents are unique
  • The causal states as defined above will give the noiseless weak latents
  • Not all systems are easily abstractable. Consider a multivariable gaussian distribution where the covariance matrix doesn't have a low-rank part. The covariance matrix is symmetric positive - after diagonalization the eigenvalues should be roughly equal.
  • Consider a sequence of buckets  and you put messages  in two buckets . In this case the minimal latent has to remember all the messages - so the latent is large. On the other hand, we can quotient : all variables become independent. 

EDIT: Sam Eisenstat pointed out to me that this doesn't work. The construction actually won't satisfy the 'stability criterion'.

The noiseless natural latent might not always exist. Indeed consider a generic distribution    on . In this case, the causal state cosntruction will just yield a copy of . In this case the reconstructavility/stability criterion is not satisfied. 

Replies from: alexander-gietelink-oldenziel, johnswentworth
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-08-03T16:31:13.894Z · LW(p) · GW(p)

Inspired by this Shalizi paper defining local causal states. The idea is so simple and elegant I'm surprised I had never seen it before. 

Basically, starting with a a factored probability distribution  over a dynamical DAG  we can use Crutchfield causal state construction locally to construct a derived causal model factored over  the dynamical DAG as  where  is defined by considering the past and forward lightcone of  defined as  all those points/ variables  which influence  respectively are influenced by  (in a causal interventional sense) . Now take define the equivalence relatio on realization  of   (which includes  by definition)[1] whenever the conditional probability distribution   on the future light cones are equal. 

These factored probability distributions over dynamical DAGs are called 'fields' by physicists. Given any field  we define a derived local causal state field  in the above way. Woah!

 

Some thoughts and questions

  • this depends on the choice of causal factorizations. Sometimes these causal factorizations are given but in full generality one probably has to consider all factorizations simultaneously, each giving a  different local state presentation!
    • What is the Factored sets angle here?
    • In particular, given a stochastic process  the reverse  can give a wildly different local causal field as minimal predictors and retrodictors can be different. This can be exhibited by the random insertion process, see this  paper.  
  • Let a stochastic process  be given and define the (forward) causal states  as usual. The key 'stochastic complexity' quantity is defined as the mutual information  of the causal states and the past. We may generalize this definition, replacing the past with the local past lightcone to give a local stochastic complexity. 
    • Under the assumption that the stochastic process is ergodic the causal state form an irreducible Hidden Markov Model and the stochastic complexity can be calculated as the entropy of the stationary distribution. 
    • !!Importantly, the stochastic complexity is different from the 'excess entropy' of the mutual information of the past (lightcone) and the future (lightcone). 
    • This gives potentially a lot of very meaningful quantities to compute. These are I think related to correlation functions but contain more information in general. 
  • Note that the local causal state construction is always possible - it works in full generality. Really quite incredible!
  • How are local causal fields related to Wentworth's latent natural abstractions? 
  • Shalizi conjectures that the local causal states form a Markov field - which would mean by Hammersley-Clifford we could describe the system as a Gibb distribution ! This would prove an equivalence between the Gibbs/MaxEnt/ Pitman-Koopman-Darmois theory and the conditional independence story of Natural Abstraction roughly similar to early approaches of John. 
    • I am not sure what the status of the conjecture is at this moment. It seems rather remarkable that such a basic fact, if true, cannot be proven. I haven't thought about it much but perhaps it is false in a subtle way. 
    • A Markov field factorizes over an undirected graph which seems strictly less general than a directed graph. I'm confused about this. 
  • Given a symmetry group  acting on the original causal model /field  the action will descend to an action  on the derived local causal state field. 
    • A stationary process  is exactly one with a translation action by . This underlies the original epsilon machine construction of Crutchfield, namely the fact that the causal states don't just form a set (+probability distribution) but are endowed with a monoid structure -> Hidden Markov Model.  

 

  1. ^

    In other words, by convention the Past includes the Present  while the Future excludes the Present. 

Replies from: Darcy
comment by Dalcy (Darcy) · 2024-07-11T16:37:41.181Z · LW(p) · GW(p)

Just finished the local causal states paper, it's pretty cool! A couple of thoughts though:

I don't think the causal states factorize over the dynamical bayes net, unlike the original random variables (by assumption). Shalizi doesn't claim this either.

  • This would require proving that each causal state is conditionally independent of its nondescendant causal states given its parents, which is a stronger theorem than what is proved in Theorem 5 (only conditionally independent of its ancestor causal states, not necessarily all the nondescendants)

Also I don't follow the Markov Field part - how would proving:

if we condition on present neighbors of the patch, as well as the parents of the patch, then we get independence of the states of all points at time t or earlier. (pg 16)

... show that the causal states is a markov field (aka satisfies markov independencies (local or pairwise or global) induced by an undirected graph)? I'm not even sure what undirected graph the causal states would be markov with respect to. Is it the ...

  • ... skeleton of the dynamical Bayes Net? that would require proving a different theorem: "if we condition on parents and children of the patch, then we get independence of all the other states" which would prove local markov independency
  • ... skeleton of the dynamical Bayes Net + edges for the original graph for each t? that would also require proving a different theorem: "if we condition on present neighbors, parents, and children of the patch, then we get independence of all the other states" which would prove local markov independency

Also for concreteness I think I need to understand its application in detecting coherent structures in cellular automata to better appreciate this construction, though the automata theory part does go a bit over my head :p

comment by johnswentworth · 2023-06-30T16:09:37.515Z · LW(p) · GW(p)

Defn. A natural latent is noiseless if ?

 ??

[Intuitively,  should contain no independent noise not accoutned for by the ]

That condition doesn't work, but here's a few alternatives which do (you can pick any one of them):

  •  - most conceptually confusing at first, but most powerful/useful once you're used to it; it's using the trick from Minimal Map [LW · GW].
  • Require that  be a deterministic function of , not just any latent variable.

(The latter two are always equivalent for any two variables  and are somewhat stronger than we need here, but they're both equivalent to the first once we've already asserted the other natural latent conditions.)

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-06-02T18:23:31.841Z · LW(p) · GW(p)

Reasons to think Lobian Cooperation is important

Usually the modal Lobian cooperation is dismissed as not relevant for real situations but it is plausible that Lobian cooperation extends far more broadly than what is proved currently.

 It is plausible that much of cooperation we see in the real world is actually approximate Lobian cooperation rather than purely given by traditional game-theoretic incentives. 
Lobian cooperation is far stronger in cases where the players resemble each other and/or have access to one another's blueprint. This is arguably only very approximately the case between different humans but it is much closer to be the case when we are considering different versions of the same human through time as well as subminds of that human. 


In the future we may very well see probabilistically checkable proof protocols, generalized notions of proof like heuristic arguments, magical cryptographic trust protocols and formal computer-checked contracts widely deployed. 

All these considerations could potentially make it possible for future AI societies to exhibit vastly more cooperative behaviour. 

Artificial minds also have several features that make them intrinsically likely to engage in Lobian cooperation. i.e. their easy copyability (which might lead to giant 'spur' clans).  Artificial minds can be copied, their source code and weight may be shared and the widespread use of simulations may become feasible. All these point towards the importance of Lobian cooperation and Open-Source Game theory more generally. 

[With benefits also come drawbacks like the increased capacity for surveillance and torture. Hopefully, future societies may develop sophisticated norms and technology [LW · GW] to avoid these outcomes. ]

The Galaxy brain take is the trans-multi-Galactic brain of Acausal Society. [LW · GW]
 

Replies from: sharmake-farah
comment by Noosphere89 (sharmake-farah) · 2023-06-02T18:38:27.838Z · LW(p) · GW(p)

I definitely agree that cooperation can definitely be way better in the future, and Lobian cooperation, especially with Payor's Lemma, might well be enough to get coordination across entire solar system.

That stated, it's much more tricky to expand this strategy to galactic scales, assuming our physical models aren't wrong, because light speed starts to become a very taut constraint under a galaxy wide brain, and acausal strategies will require a lot of compute to simulate entire civilizations. Even worse, they depend on some common structure of values, and I suspect it's impossible to do in the fully general case.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-01T11:40:27.058Z · LW(p) · GW(p)

Pseudorandom warp fields

A highly exaggerated and intensely oscillatory 1D loss landscape representing a neural network training on a pseudorandom-hard function. The landscape should feature extremely sharp, frequent peaks and valleys, showing an almost chaotic and warped pattern. Include intense fluctuations and dramatic ridges, illustrating a landscape that is incredibly difficult to optimize. The overall visual should convey an impression of a 'cursed' optimization path, with a vibrant color scheme to emphasize the oscillatory and warped nature.

[tl;dr the loss landscape around a set of weights encoding an unlearnable 'pseudorandom' function will be warped in such a way that gradient optimizers will bob around for exponentially long. ]

Unlearnable Functions: Sample Complexity and Time Complexity

Computational learning theory contains numerous 'no-go' results indicating that many functions are not tractably learnable.

The most classical result is probably the VC dimension and PAC learnability. A good example to think about are parity functions. The output is, in some sense, very (maximally) sensitive to the input*. Effectively, this means that we need a lot of data points to pin down the functions. These functions require an intractable number of data points to learn effectively due to their high sample complexity.

Using pseudorandom generators (e.g., Blum-Blum-Shub), one can similarly construct functions that need a superpolynomial amount of time to learn.

Neural networks are very expressive. The oft-cited universal approximation theorem states that any function may be approximated by neural networks (but be careful with the required network size and depth). These superpolynomial-hard functions may be expressed by neural networks, but they wouldn't be learnable due to optimization challenges and the exponential time required.

 

Gradient Optimizers and Cursed Landscapes

There is some old work by Stuart Kauffman investigating what kinds of landscapes are learnable by optimization processes like evolution by natural selection. The takeaway is that 'most' fitness landscapes aren't navigable by optimization processes akin to gradient descent; there are too many local minima, making efficient search infeasible.

I conjecture that the landscape around a set of weights w∈Ww \in Ww∈W encoding a pseudorandom-hard function fff becomes extremely oscillatory. This warping causes any gradient optimizer (regardless of step size) to get stuck in local minima or wander around for exponentially long periods***.

This has practical implications for reasoning about the inductive bias of Stochastic Gradient Descent (SGD). Most existing convergence proofs for gradient optimizers impose conditions such as Lipschitz continuity on the landscape. Or they will talk about 'long tails of ϵ\epsilonϵ-approximate saddle points'.


*This is the start of the beautiful field of Fourier analysis of Boolean circuits. Thanks to Jake and Kaarel for introducing me to this circle of ideas recently—it's very cool!

**Here's something I am confused about: the Statistical Learning Theory (SLT) point of view would be that parameters encoding these functions would have large λ\lambdaλ (e.g., eigenvalues of the Hessian). I am a little confused about this since λ\lambdaλ is bounded by d2\frac{d}{2}2d​, which can be high but not exponentially large. Perhaps this is where the mysterious multiplicity parameter mmm comes in—or lower-order contributions to the free-energy formula.

***Perhaps 'superpolynomially long' is a more accurate description.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-04-14T19:33:24.350Z · LW(p) · GW(p)

Does internal bargaining and geometric rationality explain ADHD & OCD?

Self- Rituals as Schelling loci for Self-control and OCD

Why do people engage in non-social Rituals 'self-rituals'? These are very common and can even become pathological (OCD). 

High-self control people seem to more often have OCD-like symptoms. 

One way to think about self-control is as a form of internal bargaining between internal subagents. From this perspective, Self-control, time-discounting can be seen as a resource. In the absence of self-control the superagent 
Do humans engage in self-rituals to create Schelling points for internally bargaining agents?

Exploration, self-control, internal bargaining, ADHD

Why are exploration behaviour and lack of selfcontrol linked ? As an example ADHD-people often lack self-control, conscientiousness. At the same time, they explore more. These behaviours are often linked but it's not clear why. 

It's perfectly possible to explore, deliberately. Yet, it seems that the best explorers are highly correlated with lacking self-control. How could that be?

There is a boring social reason: doing a lot of exploration often means shirking social obligations. Self-deceiving about your true desires might be the only way to avoid social repercussions. This probably explains a lot of ADHD - but not necessarily all. 

If self-control = internal bargaining then it would follow that a lack of self-control is a failure of internal bargaining. Note that with subagents I mean both subagents in space  *and*  time . From this perspective an agent through time could alternatively be seen as a series of subagents of a 4d worm superagent. 

This explains many of the salient features of ADHD:

[Claude, list salient features and explain how these are explained by the above]

  1. Impulsivity: A failure of internal subagents to reach an agreement intertemporaly, leading to actions driven by immediate desires.
  2. Difficulty with task initiation and completion: The inability of internal subagents to negotiate and commit to a course of action.
  3. Distractibility: A failure to prioritize the allocation of self-control resources to the task at hand.
  4. Hyperfocus: A temporary alignment of internal subagents' interests, leading to intense focus on engaging activities.
  5. Disorganization: A failure to establish and adhere to a coherent set of priorities across different subagents.
  6. Emotional dysregulation: A failure of internal bargaining to modulate emotional reactions.

Arithmetic vs Geometric Exploration. Entropic drift towards geometric rationality

[this section obviously owes a large intellectual debt to Garrabrant's geometric rationality sequence]

Sometimes people like to say that geometric exploration = kelly betting =maximizing geometric mean is considered to be 'better' than arithmetic mean.

The problem is that actually just maximizing expected value rather than geometric expected value does in fact maximize the total expected value, even for repeated games (duh!). So it's not really clear in what sense geometric maximization is better in a naive sense. 

Instead, Garrabrant suggests that it is better to think of geometric maximizing as a part of a broader framework of geometric rationality wherein Kelly betting, Nash bargaining, geometric expectation are all forms of cooperation between various kinds of subagents. 

If self-control is a form of sucessful internal bargaining then it is best to think of it as a  resource. It is better to maximize arithmetic mean but it means that subagents need to cooperate & trust each other much more. Arithmetic maximization means that the variance of outcomes between future copies of the agent is much larger than geometric maximization. That means that subagents should be more willing to take a loss in one world to make up for it in another.

It is hard to be coherent

It is hard to be a coherent agent. Coherence and self-control are resources. Note that having low time-discounting is also a form of coherence: it means the subagents of the 4d-worm superagent are cooperating. 

Having subagents that are more similar to one another means it will be easier for them to cooperate. Conversely, the less they are alike the harder it is to cooperate and to be coherent. 

Over time, this means there is a selective force against an arithmetic mean maximizing superagent. 

Moreover, if the environment is highly varied (for instance when the agent select the environment to be more variable because it is exploring) the outcomes for subagents is more varied so there is  more  entropic pressure on the superagent. 

This means that in particular we would expect superagents that explore more (ADHDers) are less coherent over time (higher time-discounting) and space (more internal conflict etc). 

Replies from: quetzal_rainbow
comment by quetzal_rainbow · 2024-04-14T21:08:53.390Z · LW(p) · GW(p)

I feel like the whole "subagent" framework suffers from homunculus problem: we fail to explain behavior using the abstraction of coherent agent, so we move to the abstraction of multiple coherent agents, and while it can be useful, I don't think it displays actual mechanistic truth about minds.

When I plan something and then fail to execute plan it's mostly not like "failure to bargain". It's just when I plan something I usually have good consequences of plan in my imagination and this consequences make me excited and then I start plan execution and get hit by multiple unpleasant details of reality. Coherent structure emerges from multiple not-really-agentic pieces.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-04-15T07:55:11.233Z · LW(p) · GW(p)

You are taking subagents too literally here. If you prefer take another word like shard, fragment, component, context-dependent action impulse generator etc

Replies from: quetzal_rainbow
comment by quetzal_rainbow · 2024-04-15T09:01:58.781Z · LW(p) · GW(p)

When I read word "bargaining" I assume that we are talking about entities that have preferences, action set, have beliefs about relations between actions and preferences and exchange information (modulo acausal interaction) with other entities of the same composition. Like, Kelly betting is good because it equals to Nash bargaining between versions of yourself from inside different outcomes and this is good because we assume that you in different outcomes are, actually, agent with all arrtibutes of agentic system. Saying "systems consist of parts, this parts interact and sometimes result is a horrific incoherent mess" is true, but doesn't convey much of useful information.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-12-14T18:49:41.249Z · LW(p) · GW(p)

(conversation with Scott Garrabrant)

Destructive Criticism

Sometimes you can say something isn't quite right but you can't provide an alternative.

  • rejecting the null hypothesis
  • give a (partial) countermodel that shows that certain proof methods can't prove $A$ without proving $\neg A$. 
  • Looking at Scott Garrabrant's game of life board - it's not white noise but I can't say why

Difference between 'generation of ideas' and 'filtration of ideas' - i.e. babble and prune. 

ScottG: Bayesian learning assumes we are in a babble-rich environment and only does pruning. 

ScottG: Bayesism doesn't say 'this thing is wrong' it says 'this other thing is better'. 

Alexander: Nonrealizability the Bayesian way of saying: not enough babble?

Scott G: mwah, that suggests the thing is 'generate more babble' when the real solution is 'factor out your model in pieces and see where the culprit is'.

ergo, locality is a virtue

Alexander: locality just means conditional independence? Or does it mean something more?

ScottG: loss of locality means there is existenial risk

Alexander: reminds me of Vanessa's story: 

 trapped environments aren't in general learnable. This is a problem since real life is trapped. A single human life is filled to the brim with irreversible transitions & decisions. Humanity as a whole is much more robust because of locality: it is effectively playing the human life game lots of times in parallel. The knowledge gained is then redistributed through culture and genes. This breaks down when locality breaks down -> existential risk. 

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-10-25T19:45:31.720Z · LW(p) · GW(p)

Reasonable interpretations of Recursive Self Improvement are either trivial, tautological or false?

  1. (Trivial)  AIs will do RSI by using more hardware - trivial form of RSI
  2.  (Tautological) Humans engage in a form of (R)SI when they engage in meta-cognition. i.e. therapy is plausibly a form of metacognition. Meta-cognition is  plausible one of the remaining hallmarks of true general intelligence. See Vanessa Kosoy's "Meta-Cognitive Agents". 
    In this view, AGIs will naturally engage in meta-cognition because they're generally intelligent. They may (or may) not also engage in significantly more metacognition than humans but this isn't qualitatively different from what the human cortical algorithm already engages in. 
  3. (False) It's plausible that in many domains learning algorithms are already near a physical optimum. Given a fixed Bayesian prior of prior information and a data-set the Bayesian posterior is precise formal sense the ideal update. In practice Bayesian updating is intractable so we typically sample from the posterior using something SGD. It is plausible that something like SGD is already close to the optimum for a given amount of compute.  
Replies from: Vladimir_Nesov, mtrazzi, niplav, fread2281, thomas-kwa, lcmgcd
comment by Vladimir_Nesov · 2023-10-26T02:27:42.806Z · LW(p) · GW(p)

SGD finds algorithms. Before the DL revolution, science studied such algorithms. Now, the algorithms become inference without as much as a second glance. With sufficient abundance of general intelligence brought about by AGI, interpretability might get a lot out of studying the circuits SGD discovers. Once understood, the algorithms could be put to more efficient use, instead of remaining implicit in neural nets and used for thinking together with all the noise that remains from the search.

comment by Michaël Trazzi (mtrazzi) · 2023-10-25T22:08:13.026Z · LW(p) · GW(p)

I think most interpretations of RSI aren't useful.

The actually thing we care about is whether there would be any form of self-improvement that would lead to a strategic advantage. The fact that something would "recursively" self-improve 12 times or 2 times don't really change what we care about. 

With respect to your 3 points.

1) could happen by using more hardware, but better optimization of current hardware / better architecture is the actually scary part (which could lead to the discovery of "new physics" that could enable an escape even if the sandbox was good enough for the model before a few iterations of the RSI).

2) I don't think what you're talking about in terms of meta-cognition is relevant to the main problem. Being able to look at your own hardware or source code is though.

3) Cf. what I said at the beginning. The actual "limit" is I believe much higher than the strategic advantage threshold.

comment by niplav · 2023-10-25T20:47:51.842Z · LW(p) · GW(p)

:insightful reaction:

In practice Bayesian updating is intractable so we typically sample from the posterior using something SGD. It is plausible that something like SGD is already close to the optimum for a given amount of compute.

I give this view ~20%: There's so much more info in some datapoints (curvature, third derivative of the function, momentum, see also Empirical Bayes-like SGD, the entire past trajectory through the space) that seems so available and exploitable!

comment by acertain (fread2281) · 2023-10-25T20:39:00.449Z · LW(p) · GW(p)

What about specialized algorithms for problems (e.g. planning algorithms)?

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-10-25T21:15:46.286Z · LW(p) · GW(p)

What do you mean exactly? There are definitely domains in which humans have not yet come close to optimal algorithms.

Replies from: fread2281
comment by acertain (fread2281) · 2024-11-01T07:13:20.820Z · LW(p) · GW(p)

I guess this is sorta about your 3, which I disbelieve (though algorithms for tasks other than learning are also important). Currently, Bayesian inference vs SGD is a question of how much data you have (where SGD wins except for very little data). For small to medium amounts of data, even without AGI, I expect SGD to lose eventually due to better inference algorithms. For many problems I have the intuition that it's ~always possible to improve performance with more complicated algorithms (eg sat solvers). All that together makes me expect there to be inference algorithms that scale to very large amounts of data (that aren't going to be doing full Bayesian inference but rather some complicated approximation).

comment by Thomas Kwa (thomas-kwa) · 2023-10-25T20:07:37.769Z · LW(p) · GW(p)

What about automated architecture search?

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-10-25T20:19:51.266Z · LW(p) · GW(p)

Architectures mostly don't seem to matter, see 3. 

When they do (like in Vanessa's meta-MDPs) I think it's plausible automated architecture search is a simply an instantiation of the algorithm for general intelligence (see 2.)

comment by lukehmiles (lcmgcd) · 2023-10-25T22:21:08.201Z · LW(p) · GW(p)

I think the AI will improve (itself) via better hardware and algorithms, and it will be a slog. The AI will frequently need to do narrow tasks where the general algorithm is very inefficient.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-10-25T22:37:03.291Z · LW(p) · GW(p)

As I state in the OP I don't feel these examples are nontrivial examples of RSI.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-09-23T17:13:23.205Z · LW(p) · GW(p)

Trivial but important

Aumann agreement can fail for purely epistemic reasons because real-world minds do not do Bayesian updating. Bayesian updating is intractable so realistic minds sample from the prior. This is how e.g. gradient descent works and also how human minds work.

In this situation a two minds can end in two different basins with similar loss on the data. Because of computational limitations. These minds can have genuinely different expectation for generalization.

(Of course this does not contradict the statement of the theorem which is correct.)

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-08-26T21:52:43.426Z · LW(p) · GW(p)

Imprecise Information theory 

Would like a notion of entropy for credal sets. Diffractor suggests the following:

let  be a credal set. 

Then the entropy of  is defined as

where  denotes the usual Shannon entropy.

I don't like this since it doesn't satisfy the natural desiderata below. 


Instead, I suggest the following. Let  denote the (absolute) maximum entropy distribution, i.e.  and let .

Desideratum 1: 

Desideratum 2: Let  and consider 

Then .

Remark. Check that these desiderata are compatible where they overlap.

It's easy to check that the above 'maxEnt'- suggestion satisfies these desiderata.

Entropy operationally

Entropy is really about stochastic processes more than distributions. Given a distribution  there is an associated stochastic process  where  is sampled iid from . The entropy is really about the expected code length of encoding samples from this process.

In the credal set case there are two processes that can be naturally associated with a credal set  . Basically, do you pick a  at the start and then sample according to   (this is what Diffractors entropy refers to) or do you allow the environment to 'choose' each round a different 

In the latter case, you need to pick an encoding that does least badly. 

[give more details. check that this makes sense!]

Properties of credal maxEnt entropy

We may now investigate properties of the entropy measure. 

remark. This is different from the following measure!

Remark. If we think of  as denoting the amount of bits we receive when we know that  holds and we sample from  uniformly then   denotes the number of bits we receive when find out that  when  we knew 

What about 

?

...?

we want to do an presumption of independence - mobius/ Euler characteristic expansion

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2022-12-18T12:09:06.131Z · LW(p) · GW(p)

Roko's basilisk is a thought experiment which states that an otherwise benevolent artificial superintelligence (AI) in the future would be incentivized to create a virtual reality simulation to torture anyone who knew of its potential existence but did not directly contribute to its advancement or development.

Why Roko's basilisk probably doesn't work for simulation fidelity reasons: 

Roko's basilisk threatens to simulate and torture you in the future if you don't comply. Simulation cycles cost resources. Instead of following through on torturing our would-be cthulhu worshipper they could spend those resources on something else.

 But wait can't it use acausal magic to precommit to follow through? No.

Acausal arguments only work in situations where agents can simulate each others with high fidelity. Roko's basilisk can simulate the human but not the other way around! The human's simulation of Roko's basilisk is very low fidelity - in particular Roko's Basilisk is never confused whether or not it is being simulated by a human - it knows for a fact that the human is not able to simulate it. 

I thank Jan P. for coming up with this argument. 

Replies from: Vladimir_Nesov, Richard_Kennaway, TAG
comment by Vladimir_Nesov · 2022-12-18T15:59:55.012Z · LW(p) · GW(p)

Acausal arguments only work in situations where agents can simulate each others with high fidelity.

If the agents follow simple principles [EA · GW], it's simple to simulate those principles with high fidelity, without simulating each other in all detail. The obvious guide to the principles that enable acausal coordination is common knowledge [LW(p) · GW(p)] of each other, which could be turned into a shared agent [LW(p) · GW(p)] that adjudicates a bargain on their behalf.

comment by Richard_Kennaway · 2022-12-19T08:18:49.872Z · LW(p) · GW(p)

I have always taken Roko's Basilisk to be the threat that the future intelligence will torture you, not a simulation, for not having devoted yourself to creating it.

comment by TAG · 2022-12-18T16:33:45.565Z · LW(p) · GW(p)

How do you know you are not in a low fidelity simulation right now? What could you compare it against?

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-12-14T21:26:38.215Z · LW(p) · GW(p)

All concepts can be learnt. All things worth knowing may be grasped. Eventually.

All can be understood - given enough time and effort.

For Turing-complete organism, there is no qualitive gap between knowledge and ignorance. 

No qualitive gap but one. The true qualitative difference: quantity. 

Often we simply miss a piece of data. The gap is too large - we jump and never reach the other side. A friendly hominid who has trodden the path before can share their journey. Once we know the road, there is no mystery. Only effort and time. Some hominids choose not to share their journey. We keep a special name for these singular hominids: genius.

Replies from: Viliam
comment by Viliam · 2023-12-16T14:42:00.141Z · LW(p) · GW(p)

Given enough time

Well, that's exactly the problem.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-11-20T12:14:33.023Z · LW(p) · GW(p)

Abnormalised sampling?
Probability theory talks about sampling for probability distributions, i.e. normalized measures. However, non-normalized measures abound: weighted automata, infra-stuff, uniform priors on noncompact spaces, wealth in logical-inductor esque math, quantum stuff?? etc.

Most of probability theory constructions go through just for arbitrary measures, doesn't need the normalization assumption. Except, crucially, sampling. 

What does it even mean to sample from a non-normalized measure? What is unnormalized abnormal sampling?

I don't know. 

Infra-sampling has an interpretation of sampling from a distribution made by a demonic choice. I don't have good interpretations for other unnormalized measures. 

 

Concrete question: is there a law of large numbers for unnormalized measures? 

Let f be a measureable function and m a measure. Then the expectation value is defined . A law of large numbers for unnormalized measure would have to say something about repeated abnormal sampling. 

 

I have no real ideas. Curious to learn more. 

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-09-28T20:56:53.693Z · LW(p) · GW(p)

SLT and phase transitions

The morphogenetic SLT story says that during training the Bayesian posterior concentrates around a series of subspaces  with rlcts   and losses . As the size of the data sample  is scaled the Bayesian posterior makes transitions  trading off higher complexity (higher ) for better accuracy (lower loss ).

This is the radical new framework of SLT: phase transitions happen in pure Bayesian learning as the data size is scaled. 

N.B. The phase transition story actually needs a version of SLT for the nonrealizable case despite most sources focusing solely on the realizable case! The nonrealizable case makes everything more complicated and the formulas from the realizable case have to be altered. 

We think of the local RLCT  at a parameter  as a measure of its inherent complexity. Side-stepping the subtleties with this point of view let us take a look at Watanabe's formula for the Bayesian generalization error:

 

where  is a neighborhood of the local minimum  and  is its local RLCT. In our case 

 --EH I wanted to say something here but don't think it makes sense on closer inspection

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-09-28T19:00:28.560Z · LW(p) · GW(p)

Alignment by Simulation?

I've heard this alignment plan that is a variation of 'simulate top alignment researchers' with an LLM. Usually the poor alignment researcher in question is Paul. 

This strikes me as deeply unserious and I am confused why it is having so much traction. 

That AI-assisted alignment is coming (indeed, is already here!) is undeniable. But even somewhat accurately simulating a human from textdata is a crazy sci-fi ability, probably not even physically possible. It seems to ascribe nearly magical abilities to LLMs. 

Predicting a partially observable process is fundamentally hard. Even in very easy cases there are simple cases where one can give a generative (partially observable) model with just two states (the unifilar source) that needs an infinity of states to predict optimally. In more generic cases the expectation is that this is far worse. 

Error compound over time (or continuation length). Even a tiny amount of noise would throw off simulation.

Okay maybe people just mean that GPT-N will kinda know what Paul approximately would be looking at. I think this is plausible in very broad brush strokes but it seems misleading to call this 'simulation'. 

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-09-18T13:46:55.810Z · LW(p) · GW(p)

[Edit 15/05/2024: I currently think that both forward and backward chaining paradigms are missing something important. Instead, there is something like 'side-chaining' or 'wide-chaining' where you are investigating how things are related forwardly, backwardly and sideways to make use of synergystic information ]

 

Optimal Forward-chaining versus backward-chaining.

In general, this is going to depend on the domain. In environments for which we have many expert samples and there are many existing techniques backward-chaining is key.  (i.e. deploying resources & applying best practices in business & industrial contexts)

In open-ended environments such as those arising Science, especially pre-paradigmatic fields backward-chaining and explicit plans breakdown quickly. 

 

Incremental vs Cumulative

Incremental: 90% forward chaining 10% backward chaining from an overall goal. 

Cumulative: predominantly forward chaining (~60%) with a moderate amount of backward chaining over medium lengths (30%) and only a small about of backward chaining (10%) over long lengths. 

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-09-01T23:28:50.501Z · LW(p) · GW(p)

Thin versus Thick Thinking

 

Thick: aggregate many noisy sources to make a sequential series of actions in mildly related environments, model-free RL

carnal sins: failure of prioritization / not throwing away enough information , nerdsnipes, insufficient aggegration, trusting too much in any particular model,  indecisiveness, overfitting on noise, ignoring consensus of experts/ social reality

default of the ancestral environment

CEOs, general, doctors, economist, police detective in the real world, trader

Thin: precise, systematic analysis, preferably in repeated & controlled experiments to obtain cumulative deep & modularized knowledge, model-based RL

carnal sins: ignoring clues, not going deep enough, aggregating away the signal, prematurely discarding models that don't fit naively fit the evidence, not trusting formal models enough / resorting to intuition or rule of the thumb, following consensus / building on social instead of physical reality

only possible in highly developed societies with place for cognitive specalists. 

mathematicians, software engineers, engineers, historians, police detective in fiction, quant

 

Mixture: codebreakers (spying, cryptography)

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-08-27T19:11:58.747Z · LW(p) · GW(p)

[Thanks to Vlad Firoiu for helping me]

An Attempted Derivation of the Lindy Effect
Wikipedia:

The Lindy effect (also known as Lindy's Law[1]) is a theorized phenomenon by which the future life expectancy of some non-perishable things, like a technology or an idea, is proportional to their current age.

Laplace Rule of Succesion 

What is the probability that the Sun will rise tomorrow, given that is has risen every day for 5000 years? 

Let  denote the probability that the Sun will rise tomorrow. A priori we have no information on the value of  so Laplace posits that by the principle of insufficient reason one should assume a uniform prior probability  [1]

Assume now that we have observed  days, on each of which the Sun has risen.

Each event is a Bernoulli random variable  which can each be 1 (the Sun rises) or 0 (the Sun does not rise). Assume that the probability is conditionally independent of 

The likelihood of  out of  succeses according to the hypothesis   is . Now use Bayes rule 

to calculate the posterior.

Then the probability of succes for 
                                                               

This is Laplace's rule of succcesion. 

We now adapt the above method to derive Lindy's Law. 

The probability of rising  days and not rising on the  day given that the Sun rose  days is

                             The expectation of lifetime is then the average

which almost converges :o.... 

[What's the mistake here?]

 

  1. ^

    For simplicity I will exclude the cases that , see the wikipedia page for the case where they are not excluded. 

Replies from: JBlack
comment by JBlack · 2023-08-28T01:08:58.900Z · LW(p) · GW(p)

I haven't checked the derivation in detail, but the final result is correct. If you have a random family of geometric distributions, and the density around zero of the decay rates doesn't go to zero, then the expected lifetime is infinite. All of the quantiles (e.g. median or 99%-ile) are still finite though, and do depend upon n in a reasonable way.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-08-03T12:56:17.919Z · LW(p) · GW(p)

Generalized Jeffrey Prior for singular models?

For singular models the Jeffrey Prior is not well-behaved for the simple fact that it will be zero at minima of the loss function. 
Does this mean the Jeffrey prior is only of interest in regular models? I beg to differ. 

Usually the Jeffrey prior is derived as parameterization invariant prior. There is another way of thinking about the Jeffrey prior as arising from an 'indistinguishability prior'.

The argument is delightfully simple: given two weights  if they encode the same distribution  our prior weights on them should be intuitively the same . Two weights encoding the same distributions means the model exhibit non-identifiability making it non-regular (hence singular). However, regular models exhibit 'approximate non-identifiability'.

For a given dataset  of size  from the true distribution , error  we can have a whole set of weights  where the probability that  does more than  better on the loss on  than  is less than .

In other words, the sets of weights that are probabily approximately indistinguishable. Intuitively, we should assign an (approximately) uniform prior on these approximately indistinguishable regions. This gives strong constraints on the possible prior. 

The downside of this is that it requires us to know the true distribution . Instead of seeing if  are approximately indistinguishable when sampling from  we can ask if  is approximately indistinguishable from  when sampling from . For regular models this also leads to the Jeffrey prior, see this paper.  

However, the Jeffrey prior is just an approximation of this prior. We could also straightforwardly see what the exact prior is to obtain something that might work for singular models. 

 

EDIT: Another approach to generalizing the Jeffrey prior might be by following an MDL optimal coding argument - see this paper

 

Replies from: dmurfet
comment by Daniel Murfet (dmurfet) · 2023-11-27T18:32:48.120Z · LW(p) · GW(p)

You might reconstruct your sacred Jeffries prior with a more refined notion of model identity, which incorporates derivatives (jets on the geometric/statistical side and more of the algorithm behind the model on the logical side).

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-12-17T16:41:54.651Z · LW(p) · GW(p)

Is this the jet prior I've been hearing about?

I argued above that given two weights  such that they have (approximately) the same conditional distribution  the 'natural' or 'canonical' prior should assign them equal prior weights . A more sophisticated version of this idea is used to argue for the Jeffrey prior as a canonical prior. 

Some further thoughts:

  • imposing this uniformity condition would actually contradict some version of Occam's razor. Indeed,  could be algorithmically much more complex (i.e. have much higher description length) than  but they still might have similar or the same predictions. 
    • The difference between same-on-the-nose versus similar might be very material. Two conditional probability distributions might be quite similar [a related issue here is that the KL-divergence is assymetric so similarity is a somewhat ill-defined concept], yet one intrinsically requires far more computational resources. 
      • A very simple example is the uniform distribution  and another distribution  that is a small perturbation of the uniform distribution but whose exact probabilities  have decimal expansions that have very large description length (this can be produced by adding long random strings to the binary expansion). 
      • [caution: CompMech propaganda incoming] More realistic examples do occur i.e. in finding optimal predictors of dynamical systems at the edge of chaos. See the section on 'intrinsic computation of the period-doubling cascade', p.27-28 of calculi of emergence for a classical example.  
  • Asking for the prior  to restrict to be uniform for weights  that have equal/similar conditional distributions  seems very natural but it doesn't specify how the prior should relate weights with different conditional distributions. Let's say we have two weights  with very different conditional probability distributions. Let . How should we compare the prior weights ?
    Suppose I double the number of , i.e.  where we enlarged  such that  has double the volume of  and everything else is fixed. Should we have  or should the prior weight   be larger? In the former case, the a prior weight on  should be reweighted depending on how many  there are with similar conditional probability distributions, in the latter it isn't. ( Note that this is related but distinct from the parameterization invariance condition of the Jeffery prior. )
     I can see arguments for both 
    • We could want to impose the condition that quotienting out by the relation  when to not affect the model (and thereby the prior) at all.
    • On the other hand, one could argue that the Solomonoff prior would not impose  - if one finds more programs that yield  maybe you should put higher a priori credence on .  
    • The RLCT  of the new elements in  could behave wildly different from . This suggest that the above analysis is not at the right conceptual level and one needs a more refined notion of model identity. 
  • Your comment about more refined type of model identity using jets sounds intriguing. Here is a related thought
    • In the earlier discussion with Joar Skalse there was a lot of debate around whether a prior simplicity (description length, Kolmogorov complexity according to Joar) is actually captured by the RLCT. It is possible to create examples where the RLCT and the algorithmic complexity diverge.
    • I haven't had the chance to think about this very deeply but my superficial impression is that the RLCT  is best thought of as measuring a relative model complexity between  and  rather than an absolute measure of complexity of 
    • (more thoughts about relations with MDL. too scattered, I'm going to post now)
Replies from: dmurfet
comment by Daniel Murfet (dmurfet) · 2023-12-17T22:58:59.895Z · LW(p) · GW(p)

I think there's no such thing as parameters, just processes that produce better and better approximations to parameters, and the only "real" measures of complexity have to do with the invariants that determine the costs of those processes, which in statistical learning theory are primarily geometric (somewhat tautologically, since the process of approximation is essentially a process of probing the geometry of the governing potential near the parameter).

From that point of view trying to conflate parameters  such that  is naive, because  aren't real, only processes that produce better approximations to them are real, and so the  derivatives of  which control such processes are deeply important, and those could be quite different despite  being quite similar.

So I view "local geometry matters" and "the real thing are processes approximating parameters, not parameters" as basically synonymous.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2022-11-22T13:27:22.259Z · LW(p) · GW(p)

"The links between logic and games go back a long way. If one thinks of a debate as a kind of game, then Aristotle already made the connection; his writings about syllogism are closely intertwined with his study of the aims and rules of debating. Aristotle’s viewpoint survived into the common medieval name for logic: dialectics. In the mid twentieth century Charles Hamblin revived the link between dialogue and the rules of sound reasoning, soon after Paul Lorenzen had connected dialogue to constructive foundations of logic." from the Stanford Encyclopedia of Philosophy on Logic and Games

Game Semantics 

Usual presentation of game semantics of logic: we have a particular debate / dialogue game associated to a proposition between an Proponent and Opponent and Proponent tries to prove the proposition while the Opponent tries to refute it.

A winning strategy of the Proponent corresponds to a proof of the proposition. A winning strategy of the Opponent corresponds to a proof of the negation of the proposition.

It is often assumed that either the Proponent has a winning strategy in A or the Opponent has a winning strategy in A - a version of excluded middle. At this point our intuitionistic alarm bells should be ringing: we cant just deduce a proof of the negation from the absence of a proof of A. (Absence of evidence is not evidence of absence!)

We could have a situation that neither the Proponent or the Opponent has a winning strategy! In other words neither A or not A is derivable. 

Countermodels

One way to substantiate this is by giving an explicit counter model  in which  respectively  don't hold.

Game-theoretically a counter model  should correspond to some sort of strategy! It is like an "interrogation" /attack strategy that defeats all putative winning strategies. A 'defeating' strategy or 'scorched earth'-strategy if you'd like. A countermodel is an infinite strategy. Some work in this direction has already been done[1][2]

Dualities in Dialogue and Logic

This gives an additional symmetry in the system, a syntax-semantic duality distinct to the usual negation duality. In terms of proof turnstile we have the quadruple

 meaning  is provable

 meaning $$ is provable

 meaning  is not provable because there is a countermodel  where  doesn't hold - i.e. classically  is satisfiable.

 meaning  is not provable because there is a countermodel  where  doesn't hold - i.e. classically  is satisfiable.

Obligationes, Positio, Dubitatio

In the medieval Scholastic tradition of logic there were two distinct types of logic games ("Obligationes) - one in which the objective was to defend a proposition against an adversary ("Positio") the other the objective was to defend the doubtfulness of a proposition ("Dubitatio").[3]

Winning strategies in the former corresponds to proofs while winning (defeating!) strategies in the latter correspond to countermodels. 

Destructive Criticism

If we think of argumentation theory / debate a counter model strategy is like "destructive criticism" it defeats attempts to buttress evidence for a claim but presents no viable alternative. 

  1. ^

     Ludics & completeness - https://arxiv.org/pdf/1011.1625.pdf

  2. ^

    Model construction games, Chap 16 of Logic and Games van Benthem

  3. ^

    Dubitatio games in medieval scholastic tradition, 4.3 of https://apcz.umk.pl/LLP/article/view/LLP.2012.020/778

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2022-11-18T13:35:43.973Z · LW(p) · GW(p)

Ambiguous Counterfactuals

[Thanks to Matthias Georg Mayer for pointing me towards ambiguous counterfactuals]

Salary is a function of eXperience and Education

We have a candidate  with given salary, experience  and education .

Their current salary is given by 


We 'd like to consider the counterfactual where they didn't have the education . How do we evaluate their salary in this counterfactual?

This is slightly ambiguous - there are two counterfactuals:

 or  

In the second counterfactual, we implicitly had an additional constraint , representing the assumption that the candidate would have spent their time either in education or working. Of course, in the real world they could also have dizzled their time away playing video games.

One can imagine that there is an additional variable: do they live in a poor country or a rich country. In a poor country if you didn't go to school you have to work. In a rich country you'd just waste it on playing video games or whatever. Informally, we feel in given situations one of the counterfactuals is more reasonable than the other. 

Coarse-graining and Mixtures of Counterfactuals

We can also think of this from a renormalization / coarsegraining story. Suppose we have a (mix of) causal models coarsegraining a (mix of) causal models. At the bottom we have the (mix of? Ising models!) causal model of physics. i.e. in electromagnetics the Green functions give use the intervention responses to adding sources to the field.

A given counterfactual at the macrolevel can now have many different counterfactuals at the microlevels. This means we actually would get a probability dsitribution of likely counterfactuals at the top levels. i.e. in 1/3 of the cases the candidate actually worked the 5 years they didn't go to school. In 2/3 of the cases the candidate just wasted it on playing video games. 

The outcome of the counterfactual  is then not a single number but a distribution

 

where  is random variable with distribution the Bernoulli distribution with bias .

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2022-11-16T15:59:55.020Z · LW(p) · GW(p)

Insights as Islands of Abductive Percolation?

I've been fascinated by this beautiful paper by Viteri & DeDeo. 

What is a mathematical insight? We feel intuitively that proving a difficult theorem requires discovering one or more key insights. Before we get into what the Dedeo-Viteri paper has to say about (mathematical) insights let me recall some basic observations on the nature of insights:

(see also my previous shortform)

  • There might be a unique decomposition, akin to prime factorization. Alternatively, there might many roads to Rome: some theorems can be proved in many different ways. 
  • There are often many ways to phrase an essentially similar insight. These different ways to name things we feel are 'inessential'. Different labelings should be easily convertible into one another. 
  • By looping over all possible programs all proofs can be eventually found, so the notion of an 'insight' has to fundamentally be about feasibility.
  • Previously, I suggested a required insight is something like a private key to a trapdoor function. Without the insight you are facing an infeasible large task. With it, you can suddenly easily solve a whole host of new tasks/ problems
  • Insight may be combined in (arbitrarily?) complex ways. 

 When are two proofs of essentially different?

Some theorems can be proved in many different ways. That is  different in the informal sense. It isn't immediately clear how to make this more precise.

We could imagine there is a whole 'homotopy' theory of proofs, but before we do so we need to understand when two proofs are essentially the same or essentially different. 

  • On one end of the spectrum, proofs can just be syntactically different but we feel they have 'the same content'. 
  • We can think type-theoretically, and say two proofs are the same when their denotations (normal forms) are the same. This is obviously better than just asking for syntactical equality or apartness. It does mean we'd like some sort of intuitionistic/type-theoretic foundation since a naive classicial foundations makes all normals forms equivalent. 
  • We can also look at what assumptions are made in the proof. I.e. one of the proofs might use the Axiom of Choice, while the other does not. An example is the famous nonconstructive proof of the irrationality of  which turns out to have a constructive proof as well. 
  • If we consider proofs as functorial algorithms we can use mono-Anabelian transport to distinguish them in some case. [LINK!]
  • We can also think homotopy type-theoretically and ask when two terms of a type are equal in the HoTT sense. 

With the exception of the mono-anabelian transport one - all these suggestions of 'don't go deep enough', they're too superficial.

Phase transitions and insights, Hopfield Networks & Ising Models

(See also my shortform on Hopfield Networks/ Ising models as mixtures of causal models [LW(p) · GW(p)])

Modern ML models famously show some sort of phase transitions in understanding. People have been especially fascinated by the phenomenon of 'grokking, see e.g. here and here. It suggests we think of insights in terms of phase transitions, critical points etc.

Dedeo & Viteri have an ingenious variation on this idea. They consider a collection of famous theorems and their proofs formalized in a proof assistant. 

They then imagine these proofs as a giant directed graph and consider a Boltzmann distributions on it. (so we are really dealing with an Ising model/ Hopfield network here). We think of this distribution as a measure of 'trust' both trust in propositions (nodes) and inferences (edges). 

We show that the epistemic relationship between claims in a mathematical proof has a network structure that enables what we refer to as an epistemic phase transition (EPT): informally, while the truth of any particular path of argument connecting two points decays exponentially in force, the number of distinct paths increases. Depending on the network structure, the number of distinct paths may itself increase exponentially, leading to a balance point where influence can propagate
at arbitrary distance (Stanley, 1971). Mathematical proofs have the structure necessary to make this possible.
In the presence of bidirectional inference—i.e., both deductive and abductive reasoning—an EPT enables a proof to produce near-unity levels of certainty even in the presence of skepticism about the validity of any particular step. Deductive and abductive reasoning, as we show, must be well-balanced for this to happen. A relative over-confidence in one over the other can frustrate the effect, a phenomenon we refer to as the abductive paradox

The proofs of these famous theorems break up into 'abductive islands'. They have natural modularity structure into lemmas. 

EPTs are a double-edged sword, however, because disbelief can propagate just as easily as truth. A second prediction of the model is that this difficulty—the explosive spread of skepticism—can be ameliorated when the proof is made of modules: groups of claims that are significantly more tightly linked to each other than to the rest of the network. 

(...)
When modular structure is present, the certainty of any claim within a cluster is reasonably isolated from the failure of nodes outside that cluster.

One could hypothesize that insights might correspond somehow to these islands.

Final thoughts

I like the idea that a mathemathetical insight might be something like an island of deductively & abductively tightly clustered propositions. 

Some questions:

  • How does this fit into the 'Natural Abstraction' - especially sufficient statistics?
  • How does this interact with Schmidthuber's Powerplay?

EDIT: The separation property of Ludics, see e.g. here, points towards the point of view that proofs can be distinguished exactly by suitable (counter)models. 

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-01T10:30:19.412Z · LW(p) · GW(p)

The pseudorandom lie under the Lava lamp

Our observations are compatible with a world that is generated by a Turing machine with just a couple thousand bits.

That means that all the seemingly random bits we see in Geiger counters, Lava lamps, gasses and the like is only pseudorandomness in actuality.

Replies from: TsviBT, Lblack, sharmake-farah, Viliam, mikhail-samin, Signer
comment by TsviBT · 2024-10-01T11:03:37.586Z · LW(p) · GW(p)

IDK why you think that TM is simpler than one that computes, say, QM. But either way, I don't know why to favor (in terms of ascribing reality-juice) worlds that are simple TMs but not worlds that are simple physics equations. You can complain that you don't know how to execute physics equations, but I can also complain that I don't know how to execute state transitions. (Presumably there's still something central and real about some things being more executable than others; I'm just saying it's not clear what that is and how it relates to reality-juice and TMs vs physics.)

Replies from: Lblack, alexander-gietelink-oldenziel
comment by Lucius Bushnaq (Lblack) · 2024-10-05T17:15:07.964Z · LW(p) · GW(p)

You can complain that you don't know how to execute physics equations

I'm confused, in what sense don't we know how to do this? Lattice quantum field theory simulations work fine. 

Replies from: TsviBT
comment by TsviBT · 2024-10-05T18:56:10.708Z · LW(p) · GW(p)

For example, we couldn't execute continuum models.

Replies from: sharmake-farah
comment by Noosphere89 (sharmake-farah) · 2024-10-05T18:59:27.645Z · LW(p) · GW(p)

Of course, just because we can't execute continuum models, or models of physics that require actually infinite computation, not just unlimited amounts of compute, doesn't mean the universe can't execute such a program.

Replies from: TsviBT
comment by TsviBT · 2024-10-05T19:16:06.334Z · LW(p) · GW(p)

Ok, another example is that physical laws are generally descriptive, not fully specified worlds. You can "simulate" the ideal gas law or Maxwell's equations but you're doing extra work beyond just what the equations say (like, you have to run "import diffeq" first, and pick a space topology, and pick EM fields) and it's not a full world.

Replies from: sharmake-farah
comment by Noosphere89 (sharmake-farah) · 2024-10-05T19:23:07.074Z · LW(p) · GW(p)

Yes, which is why I explicitly said that the scenario involves actual/manifest infinity of compute to actually implement the equations to actually make it a full world, and if you wanted to analogize physical laws to a computer system, I'd argue that they are analogous to the source code of a computer, or the rules/state of a Turing Machine, and I'm arguing that there is a very vast difference between us simulating Maxwell's equations or the ideal gas law and the universe simulating whatever physical laws we turn out to actually have, and all of the difference is the universe has an actual infinity/manifest infinity of compute like FLOPs/FLOP/s and memory such that you can actually run the equations directly without relying on shortcuts to make the problem more tractable, whereas we have to rely on shortcuts that change the physics a little but get us a reasonable answer in a reasonable time.

Replies from: TsviBT
comment by TsviBT · 2024-10-05T19:28:33.961Z · LW(p) · GW(p)

Oh I misparsed your comment somehow, I don't even remember how.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-01T12:04:38.617Z · LW(p) · GW(p)

This distinction isnt material. The distinction I am getting at is whether our physics (simulation) is using a large K-incompressible seed or not.

Replies from: TsviBT
comment by TsviBT · 2024-10-01T12:17:14.159Z · LW(p) · GW(p)

QM doesn't need a random seed!

comment by Lucius Bushnaq (Lblack) · 2024-10-05T17:12:35.248Z · LW(p) · GW(p)

The randomness of the Geiger counter comes from wave function decoherence. From the perspective of any observers who are part of the world generated by the Turing machine, this is irreducible indexical uncertainty. 

I don't know how many of the random bits in Lava lamps come from decoherence.  

comment by Noosphere89 (sharmake-farah) · 2024-10-05T19:01:55.192Z · LW(p) · GW(p)

I'm fairly sure it isn't actually compatible with a world that is generated by a Turing Machine, but the basic problem is all the real number constants in the universe which in QM are infinitely precise, not just arbitrarily precise, which wreaks havoc on Turing Machine models, but Signer has another explanation of another problem that is fatal to the approach.

comment by Viliam · 2024-10-01T12:13:15.835Z · LW(p) · GW(p)

Connotationally, even if things are pseudorandom, they still might be "random" for all practical purposes, e.g. if the only way to calculate them is to simulate the entire universe. In other words, we may be unable to exploit the pseudorandomness.

Replies from: alexander-gietelink-oldenziel
comment by Mikhail Samin (mikhail-samin) · 2024-10-01T10:46:31.933Z · LW(p) · GW(p)
  • Probability is in the mind. There's no way to achieve entanglement between what's necessary to make these predictions and the state of your brain, so for you, some of these are random.
  • In multi-worlds, the Turing machine will compute many copies of you, and there might be more of those who see one thing when they open their eyes than of those who see another thing. When you open your eyes, there's some probability of being a copy that sees one thing and a copy that sees the other thing. In a deterministic world with many copies of you, there's "true" randomness in where you end up opening your eyes.
Replies from: TsviBT
comment by TsviBT · 2024-10-01T11:00:20.593Z · LW(p) · GW(p)

I think he's saying that there's a simple-ish deterministic machine that uses pseudorandomness to make a world observationally equivalent to ours. Since it's simple, it has a lot of the reality-juice, so it's most of "where we really are".

comment by Signer · 2024-10-01T12:41:37.643Z · LW(p) · GW(p)

Our observations are compatible with a world that is generated by a Turing machine with just a couple thousand bits.

Yes, but this is kinda incompatible with QM without mangled worlds.

Replies from: alexander-gietelink-oldenziel
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2024-10-01T16:21:53.914Z · LW(p) · GW(p)

Oh ? What do you mean !

I don't know about mangled worlds

Replies from: Signer
comment by Signer · 2024-10-01T18:30:18.642Z · LW(p) · GW(p)

https://mason.gmu.edu/~rhanson/mangledworlds.html

I mean that if turing machine is computing universe according to the laws of quantum mechanics, observers in such universe would be distributed uniformly, not by Born probability. So you either need some modification to current physics, such as mangled worlds, or you can postulate that Born probabilities are truly random.

Replies from: TAG
comment by TAG · 2024-10-01T18:56:44.438Z · LW(p) · GW(p)

I mean that if turing machine is computing universe according to the laws of quantum mechanics,

I assume you mean the laws of QM except the collapse postulate.

observers in such universe would be distributed uniformly,

Not at all. The problem is that their observations would mostly not be in a classical basis.

not by Born probability.

Born probability relates to observations, not observers.

So you either need some modification to current physics, such as mangled worlds,

Or collapse. Mangled worlds is kind of a nothing burger--its a variation on the idea than interference between superposed states leads to both a classical basi and the Born probabilities, which is an old idea, but wihtout making it any more quantiative.

or you can postulate that Born probabilities are truly random.

??

Replies from: Signer
comment by Signer · 2024-10-01T19:50:07.740Z · LW(p) · GW(p)

Not at all. The problem is that their observations would mostly not be in a classical basis.

I phrased it badly, but what I mean is that there is a simulation of Hilbert space, where some regions contain patterns that can be interpreted as observers observing something, and if you count them by similarity, you won't get counts consistent with Born measure of these patterns. I don't think basis matters in this model, if you change basis for observer, observations and similarity threshold simultaneously? Change of basis would just rotate or scale patterns, without changing how many distinct observers you can interpret them as, right?

??

Collapse or reality fluid. The point of mangled worlds or some other modification is to evade postulating probabilities on the level of physics.

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-01-02T11:48:23.088Z · LW(p) · GW(p)

Evidence Manipulation and Legal Admissible Evidence

[This was inspired by Kokotaljo's shortform [LW · GW] on comparing strong with weak evidence] 


In the real world the weight of many pieces of weak evidence is not always comparable to a single piece of strong evidence. The important variable here is not strong versus weak per se but the source of the evidence. Some sources of evidence are easier to manipulate in various ways. Evidence manipulation, either consciously or emergently, is common and a large obstactle to truth-finding. 

Consider aggregating many (potentially biased) sources of evidence versus direct observation. These are not directly comparable and in many cases we feel direct observation should prevail.

This is especially poignant in the court of law: the very strict laws arounding presenting evidence are a culturally evolved mechanism to defend against evidence manipulation. Evidence manipulation may be easier for weaker pieces of evidence - see the prohibition against hearsay in legal contexts for instance.

It is occasionally suggested that the court of law should do more probabilistic and Bayesian type of reasoning. One reason courts refuse to do so (apart from more Hansonian reasons around elites cultivating conflict suppression) is that naive Bayesian reasoning is extremely susceptible to evidence manipulation. 

Replies from: ChristianKl
comment by ChristianKl · 2023-01-02T15:20:51.455Z · LW(p) · GW(p)

These are not directly comparable and in many cases we feel direct observation should prevail.

In other cases like medicine, many people argue that direct observation should be ignored ;)

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2022-11-22T14:42:24.624Z · LW(p) · GW(p)

Imagine a data stream 

 

assumed infinite in both directions for simplicity. Here  represents the current state ( the "present") and while  and  represents the future

Predictible Information versus Predictive Information

Predictible information is the maximal information (in bits) that you can derive about the future given the access to the past. Predictive information is the amount of bits that you need from the past to make that optimal prediction.

Suppose you are faced with the question of whether to buy, hold or sell Apple. There are three options so maximally  bits of information - not all of that information might be in contained in the past, there a certain part of irreductible uncertainty (entropy) about the future no matter how well you can infer the past. Think about a freak event & blacks swans like pandemics, wars, unforeseen technological breakthroughs, just cumulative aggregated noise in consumer preference etc. Suppose that irreducible uncertainty is half of  leaving us with  of (theoretically) predictible information.

To a certain degree, it might be predictible in theory to what degree buying Apple stock is a good idea. To do so, you may need to know many things about the past: Apple's earning records, position of competitiors, general trends of the economy, understanding of the underlying technology & supply chains etc. The total sum of this information is far larger than 

To actually do well on the stock market you additionally need to do this better than the competititon - a difficult task! The predictible information is quite small compared to the predictive information

Note that predictive information is always greater than predictible information: you need to at least  bits from the past to predict  bits of the future. Often it is much larger. 

Mathematical details

Predictible information is also called 'apparent stored information' or commonly 'excess entropy'.

It is defined as the mutual information  between the future and the past.

The predictive information is more difficult to define. It is also called the 'statistical complexity' or 'forecasting complexity' and is defined as the entropy of the steady equilibrium state of the 'epsilon machine' of the process. 

What is the Epsilon Machine of the process ? Define the causal states as the process as the partition on the sets of possible pasts  where two pasts  are in the same part / equivalence class when the future conditioned on  respectively is the same.

That is . Without going into too much more detail the forecasting complexity measures the size of this creature. 

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2022-11-18T14:37:14.041Z · LW(p) · GW(p)

Agent Foundations Reading List [Living Document]
This is a stub for a living document on a reading list for Agent Foundations. 

Causality

Book of Why, Causality - Pearl

Probability theory 
Logic of Science - Jaynes

comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2022-11-16T16:07:27.590Z · LW(p) · GW(p)

Hopfield Networks = Ising Models = Distributions over Causal models?

Given a joint probability distributions  famously there might be many 'Markov' factorizations. Each corresponds with a different causal model.

Instead of choosing a particular one we might have a distribution of beliefs over these different causal models. This feels basically like a Hopfield Network/ Ising Model. 

You have a distribution over nodes and an 'interaction' distribution over edges. 

The distribution over nodes corresponds to the joint probability distribution while the distribution over edges corresponds to a mixture of causal models where a normal DAG graphical causal G model corresponds to the Ising model/ Hopfield network which assigns 1 to an edge  if the edge is in G and 0 otherwise.