Posts

The new ruling philosophy regarding AI 2024-11-11T13:28:24.476Z
First and Last Questions for GPT-5* 2023-11-24T05:03:04.371Z
The national security dimension of OpenAI's leadership struggle 2023-11-20T23:57:12.491Z
Bruce Sterling on the AI mania of 2023 2023-06-29T05:00:18.326Z
Mitchell_Porter's Shortform 2023-06-01T11:45:58.622Z
ChatGPT (May 2023) on Designing Friendly Superintelligence 2023-05-24T10:47:16.325Z
How is AI governed and regulated, around the world? 2023-03-30T15:36:55.987Z
A crisis for online communication: bots and bot users will overrun the Internet? 2022-12-11T21:11:46.964Z
One night, without sleep 2018-08-16T17:50:06.036Z
Anthropics and a cosmic immune system 2013-07-28T09:07:19.427Z
Living in the shadow of superintelligence 2013-06-24T12:06:18.614Z
The ongoing transformation of quantum field theory 2012-12-29T09:45:55.580Z
Call for a Friendly AI channel on freenode 2012-12-10T23:27:08.618Z
FAI, FIA, and singularity politics 2012-11-08T17:11:10.674Z
Ambitious utilitarians must concern themselves with death 2012-10-25T10:41:41.269Z
Thinking soberly about the context and consequences of Friendly AI 2012-10-16T04:33:52.859Z
Debugging the Quantum Physics Sequence 2012-09-05T15:55:53.054Z
Friendly AI and the limits of computational epistemology 2012-08-08T13:16:27.269Z
Two books by Celia Green 2012-07-13T08:43:11.468Z
Extrapolating values without outsourcing 2012-04-27T06:39:20.840Z
A singularity scenario 2012-03-17T12:47:17.808Z
Is causal decision theory plus self-modification enough? 2012-03-10T08:04:10.891Z
One last roll of the dice 2012-02-03T01:59:56.996Z
State your physical account of experienced color 2012-02-01T07:00:39.913Z
Does functionalism imply dualism? 2012-01-31T03:43:51.973Z
Personal research update 2012-01-29T09:32:30.423Z
Utopian hope versus reality 2012-01-11T12:55:45.959Z
On Leverage Research's plan for an optimal world 2012-01-10T09:49:40.086Z
Problems of the Deutsch-Wallace version of Many Worlds 2011-12-16T06:55:55.479Z
A case study in fooling oneself 2011-12-15T05:25:52.981Z
What a practical plan for Friendly AI looks like 2011-08-20T09:50:23.686Z
Rationality, Singularity, Method, and the Mainstream 2011-03-22T12:06:16.404Z
Who are these spammers? 2011-01-20T09:18:10.037Z
Let's make a deal 2010-09-23T00:59:43.666Z
Positioning oneself to make a difference 2010-08-18T23:54:38.901Z
Consciousness 2010-01-08T12:18:39.776Z
How to think like a quantum monadologist 2009-10-15T09:37:33.643Z
How to get that Friendly Singularity: a minority view 2009-10-10T10:56:46.960Z
Why Many-Worlds Is Not The Rationally Favored Interpretation 2009-09-29T05:22:48.366Z

Comments

Comment by Mitchell_Porter on yanni's Shortform · 2025-01-17T02:08:51.940Z · LW · GW

This comment has been on my mind a lot the past week - not because I'm not ambitious, but because I've always been ambitious (intellectually at least) and frustrated in my ambitions. I've always had goals that I thought were important and neglected, I always directly pursued them from a socially marginal position rather than trying to make money first (or whatever people do when they put off their real ambitions), but I can't say I ever had a decisive breakthrough, certainly not to recognition. So I only have partial progress on a scattered smorgasbord of unfulfilled agendas, and meanwhile, after OpenAI's "o3 Christmas" and the imminent inauguration of an e/acc administration in the USA, it looks more than ever that we are out of time. I would be deeply unsurprised if it's all over by the end of the year. 

I'm left with choices like (1) concentrate on family in the final months (2) patch together what I have and use AI to quickly make the best of it (3) throw myself into AI safety. In practice they overlap, I'm doing all three, but there are tensions between them, and I feel the frustration of being badly positioned while also thinking I have no time for the meta-task of improving my position. 

Comment by Mitchell_Porter on O O's Shortform · 2025-01-15T07:45:55.170Z · LW · GW

You first might want to distinguish between national AI projects that are just about boosting the AI economy or managing the use of AI within government, and government-backed research which is specifically aimed at the AGI frontier. Presumably it's the latter that you're talking about. 

There is also the question of what a government would think it was doing, in embarking on such a project. The commercial enterprise of creating AI is already haunted by the idea that it would be bad for business if your creation wiped out the human race. That hasn't stopped anyone, but the fear is there, overcome only by greed. 

Now, what about politicians and public servants, generals and spymasters? How would they feel about leading a race to create AI? What would they think they were doing? Creating artificial super-scientists, super-soldiers, super-strategists? Compared to Silcon Valley, these people are more about the power motive than the profit motive. What, apart from the arms race, do they have to lure them along the AI path, comparable to the dream of uber-wealth that drives the tech oligarchs? (In dictatorships, I suppose there is also the dream of absolute personal power to motivate them.) 

Apart from the arms race, the vision that seems to animate pro-AI western elites, is economic and strategic competition among nations. If China takes the lead in AI, it will have the best products and the best technologies and it will conquer the world that way. So I guess the thinking of Trump 2.0's AI czar David Sacks (a friend of Thiel and Musk), and the people around him, is going to be some mixture of these themes - the US must lead because AI is the key to economic, technological, and military superiority in the 21st century. 

Now I think that even the most self-confident, gung-ho, born-to-rule man-of-destiny who gets involved in the AI race, is surely going to have a moment when they think, am I just creating my own replacement here? Can even my intellect, and my charisma, and my billions, and my social capital, really compete with something smarter than me, and a thousand times faster than me, and capable of putting any kind of human face on its activities? 

I'm not saying they're going to have a come-to-Yudkowsky moment and realize, holy crap, we'd better shut this down after all. Their Darwinist instincts will tell them that if they don't create AI first, someone else will. But perhaps they will want to be reassured. And this may be one area where techies similar to Ilya Sutskever, and Yann Lecun, and Alec Radford - i.e. the technical leads in these frontier AI research programs - may have a role in addition to their official role as chief of R&D. 

The technical people have their own dreams about what a world of AGI and ASI could look like too. They may have a story about prosperity and human flourishing with AI friends and partners. Or maybe they have a story just for their CEO masters, that even the most powerful AI, if properly trained, will just be 100% an extension of their own existing will. And who knows what kind of transhuman dreams they entertain privately, as well? 

These days, there's even the possibility that the AI itself is whispering to the corporate, political, and military leadership, telling them what they want to hear... 

I am very much speculating here, I have no personal experience of, or access to, these highest levels of power. But the psychology and ideology of the "decision-makers" - who really just seem to be riding the tiger of technical progress at this point - is surely an important feature of any such AGI Manhattan Project, too. 

Comment by Mitchell_Porter on Why do futurists care about the culture war? · 2025-01-14T15:56:06.745Z · LW · GW

Regarding Musk and Thiel, foremost they are billionaire capitalists, individuals who built enormous business empires. Even if we assume your thinking about the future is correct, we shouldn't assume that they have reproduced every step of it. You may simply be more advanced in your thinking about the future than they are. Their thought about the future crystallized in the 1980s, when they were young. Since then they have been preoccupied with building their empires. 

This raises the question, how do they see the future, and their relationship to it? I think Musk's life purpose is the colonization of Mars, so that humanity's fate isn't tied to what happens on Earth. Everything else is subordinate to that, and even robots and AI are just servants and companions for humanity in its quest for other worlds. As for Thiel, I have less sense of the gestalt of his business activities, but philosophically, the culture war seems very important to him. He may have a European sense of how self-absorbed cultural elites can narrow a nation's horizons, that drives his sponsorship of "heterodox" intellectuals outside the academy. 

If I'm right, the core of Musk's futurism is space colonization, and the core of Thiel's futurism is preserving an open society. They don't have the idea of an intelligence singularity whose outcome determines everything afterwards. In this regard, they're closer to e/acc than singularity thinking, because e/acc believes in a future that always remains open, uncertain, and pluralist, whereas singularity thinking tends towards a single apocalyptic moment in which superintelligence is achieved and irreversibly shapes the world. 

There are other reasons I can see why they would involve themselves in the culture war. They don't want a socialism that would interfere with their empires; they think (or may have thought until the last few years) that superintelligence is decades away; they see their culture war opponents as a threat to a free future (whether that is seen in e/acc or singularity terms), or even to the very existence of any kind of technological future society. 

But if I were to reduce it to one thing: they don't believe in models of the future according to which you get one thing right and then utopia follows, and they believe such thinking actually leads to totalitarian outcomes (where their definition of totalitarian may be, a techno-political order capable of preventing the building of a personal empire). Musk started OpenAI so Google wouldn't be the sole AI superpower; he was worried about centralization as such, not about whether they would get the value system right. Thiel gave up on MIRI's version of AI futurology years ago as a salvationist cult; I think he would actually prefer no AI to aligned AI, if the latter means alignment with a particular value system rather than alignment with what the user wants. 

Comment by Mitchell_Porter on A Novel Idea for Harnessing Magnetic Reconnection as an Energy Source · 2025-01-13T12:09:54.422Z · LW · GW

Thank you for the answer! What model was used - 4o? o1? 

Comment by Mitchell_Porter on A Novel Idea for Harnessing Magnetic Reconnection as an Energy Source · 2025-01-13T04:16:30.767Z · LW · GW

Was this idea or post AI-assisted in any way? If it was, it seems a cut above the usual AI-generated material, and I would be interested to know your methods. 

Comment by Mitchell_Porter on No one has the ball on 1500 Russian olympiad winners who've received HPMOR · 2025-01-13T04:13:53.664Z · LW · GW

Is there a good Russian-language introduction to AI alignment?

Comment by Mitchell_Porter on Human takeover might be worse than AI takeover · 2025-01-11T08:28:50.265Z · LW · GW

I think we could certainly afford to have much more discussion of this topic. The two forms of takeover are not absolutely distinct. Any humans who take over the world are going to be AI-enhanced, and all their observing, deciding, and acting will be heavily AI-mediated. On the other hand, any AI that takes over the world will be the product of human design and human training, and will start out embedded in an organization of human beings. 

Ideally, people would actually "solve ethics" and how to implement it in an AI, and we would only set superintelligence in motion having figured that out. While we still have time, we should be encouraging (and also constructively criticizing) attempts to solve those two big problems. We should also continue to think about what happens if the kind of AI that we have now or in the very near future, should acquire superintelligence. 

I agree with the author this much, that the values of our current AIs are in the right direction in various ways, and this improves the odds of a good outcome. But there are still various concerns, specific to AI takeover. What if an AI has deep values that are alien dispositions, and its humanism is simply an adaptation that will be shed once it no longer needs to get along with humans? What if there's something a bit wrong or a bit missing in the stew of values and dispositions instilled via training, system prompt, and conditioning? What if there's something a bit wrong or a bit missing in how it grounds its concepts, once it's really and truly thinking for itself? 

We might also want to think about what happens to a human brain that takes over the world via AI infrastructure. If Elon makes himself emperor of known space via Neuralink and Grok, what are the odds that his transhuman form is good, bad, or just alien and weird, in what it wants? 

Comment by Mitchell_Porter on XX by Rian Hughes: Pretentious Bullshit · 2025-01-09T01:56:47.575Z · LW · GW

I'm interested as to why?

I never heard of the book before, so I'm starting from a baseline of zero curiosity prior to this review. :-)  

But also, I'm curious about how all the pieces fit together, and what else might be banging around in those almost 1000 pages. I'm wondering if "the signal" is a callback to the metavirus from Neal Stephenson's Snow Crash, and to William Burroughs's "language is a virus from outer space" (which Stephenson was undoubtedly referencing too). I don't quite trust your grounds for dismissing various concepts mentioned as impossible, and (as you say) concepts in SF can have other kinds of value even if they are literally impossible. Finally, whatever else it is, it is clearly a work of imagination, and such works can say something about their zeitgeist, or can be weird untimely works that stand alone. 

Comment by Mitchell_Porter on XX by Rian Hughes: Pretentious Bullshit · 2025-01-08T16:26:46.685Z · LW · GW

Christoph Wetterich tried to make a shrinking universe work, though it seems to me it should result in the energy levels in atoms changing

Your review actually makes me more curious about the book. 

Comment by Mitchell_Porter on Don't fall for ontology pyramid schemes · 2025-01-08T03:12:05.103Z · LW · GW

Could you give some examples of ontology pyramid schemes? 

Also, can ontologies that are actually true, be the subject of pyramid schemes, as you define them?

Comment by Mitchell_Porter on By default, capital will matter more than ever after AGI · 2025-01-05T09:06:03.795Z · LW · GW

If I take the view that AGI has existed since late 2022 (ChatGPT), I can interpret all this as a statement about the present: capital matters more than ever, during AGI. 

After AGI is ASI, and at that point human capital is as valuable as ant capital or bear capital. 

Comment by Mitchell_Porter on Comment on "Death and the Gorgon" · 2025-01-03T08:07:03.640Z · LW · GW

There's a 2009 interview with a transhumanist Australian academic where Egan hints at some of his problems with transhumanism (even while stating elsewhere that human nature is not forever, that he expects conscious AI in his lifetime, that "universal immortality" might be a nice thing, and so forth). Evidently some of it is pure intellectual disagreement, and some of it is about not liking the psychological attitudes or subcultural politics that he sees. 

Comment by Mitchell_Porter on DeekSeek v3: The Six Million Dollar Model · 2025-01-01T05:08:08.041Z · LW · GW

Do we know any Chinese AI forums or streamers, comparable to what's on Reddit and YouTube?

Comment by Mitchell_Porter on What are the strongest arguments for very short timelines? · 2024-12-24T04:18:03.921Z · LW · GW

I have seen a poll asking "when will indefinite lifespans be possible?", and Eric Drexler answered "1967", because that was when cryonic suspension first became available. 

Similarly, I think we've had AGI at least since 2022, because even then, ChatGPT was an intelligence, and it was general, and it was artificial. 

(To deny that the AIs we have now have general intelligence, I think one would have to deny that most humans have general intelligence, too.)

So that's my main reason for very short timelines. We already crossed the crucial AGI threshold through the stupid serendipity of scaling up autocomplete, and now it's just a matter of refining the method, and attaching a few extra specialized modules. 

Comment by Mitchell_Porter on Panology · 2024-12-24T03:25:34.334Z · LW · GW

What's the difference between "panology" and "science"?

Comment by Mitchell_Porter on jacobjacob's Shortform Feed · 2024-12-22T09:26:24.546Z · LW · GW

By the start of April half the world was locked down, and Covid was the dominant factor in human affairs for the next two years or so. Do you think that issues pertaining to AI agents are going to be dominating human affairs so soon and so totally? 

Comment by Mitchell_Porter on Simulators · 2024-12-22T09:01:23.607Z · LW · GW

Hi - I would like you to explain, in rather more detail, how this entity works. It's "Claude", but presumably you have set it up in some way so that it has a persistent identity and self-knowledge beyond just being Claude? 

Comment by Mitchell_Porter on Seth Herd's Shortform · 2024-12-17T10:45:27.979Z · LW · GW

If I understand correctly, you're trying to figure out what Xi would do with the unlimited power offered by an intent-aligned ASI, or how he would react to the prospect of such, etc. 

Xi's character might matter, but I am impressed by the claim here that a competent Chinese ruler will be guided first by notions of good statecraft, with any details of personality or private life to be kept apart from their decisions and public actions. 

I'm sure that Chinese political history also offers many examples of big personalities and passionate leaders, but that would be more relevant to times when the political order is radically in flux, or needs to be rebuilt from nothing. Xi came to power within a stable system. 

So you might want to also ask how the Chinese system and ideology would respond to the idea of superintelligent AI - that is, if they are even capable of dealing with the concept! There must be considerable probability that the system would simply tune out such threatening ideas, in favor of tamer notions of AI - we already see this filtering at work in the West. 

I suppose one possibility is that they would view AI, properly employed, as a way to realize the communist ideal for real. Communist countries always say that communism is a distant goal, for now we're building socialism, and even this socialism looks a lot like capitalism these days. And one may say that the powerbrokers in such societies have long since specialized in wielding power under conditions of one-party capitalism and mercantile competition, rather than the early ideal of revolutionary leveling for the whole world. Nonetheless, the old ideal is there, just as the religious ideals still exert a cultural influence in nominally secular societies descended from a religious civilization. 

When I think about Chinese ASI, the other thing I think about, is their online fantasy novels, because that's the place in Chinese culture where they deal with scenarios like a race to achieve power over the universe. They may be about competition to acquire the magical legacy of a vanished race of immortals, rather than competition to devise the perfect problem-solving algorithm, but this is where you can find a Chinese literature that explores the politics and psychology of such a competition, all the way down to the interaction between the private and public lives of the protagonists. 

Comment by Mitchell_Porter on Mitchell_Porter's Shortform · 2024-12-09T04:42:25.069Z · LW · GW

Alexander Dugin speaks of "trumpo-futurism" and "dark accelerationism"

Dugin is a kind of Zizek of Russian multipolar geopolitical thought. He's always been good at quickly grasping new political situations and giving them his own philosophical sheen. In the past he has spoken apocalyptically of AI and transhumanism, considering them to be part of the threat to worldwide tradition coming from western liberalism. I can't see him engaging in wishful thinking like "humans and AIs coexist as equals" or "AIs migrate to outer space leaving the Earth for humans", so I will be interested to see what he says going forward. I greatly regret that his daughter (Daria Dugina) was assassinated, because she was taking a serious interest in the computer age's ontology of personhood, but from a Neoplatonist perspective; who knows what she might have come up with. 

Comment by Mitchell_Porter on Dave Kasten's AGI-by-2027 vignette · 2024-11-27T11:46:12.737Z · LW · GW

Started promisingly, but like everyone else, I don't believe in the ten-year gap from AGI to ASI. If anything, we got a kind of AGI in 2022 (with ChatGPT), and we'll get ASI by 2027, from something like your "cohort of Shannon instances". 

Comment by Mitchell_Porter on Why We Wouldn't Build Aligned AI Even If We Could · 2024-11-25T10:27:07.854Z · LW · GW

For my part, I have been wondering this week, what a constructive reply to this would be. 

I think your proposed imperatives and experiments are quite good. I hope that they are noticed and thought about. I don't think they are sufficient for correctly aligning a superintelligence, but they can be part of the process that gets us there. 

That's probably the most important thing for me to say. Anything else is just a disagreement about the nature of the world as it is now, and isn't as important. 

Comment by Mitchell_Porter on For progress to be by accumulation and not by random walk, read great books · 2024-11-25T09:40:50.887Z · LW · GW

Perhaps he means something like what Keynes said here

Comment by Mitchell_Porter on Why We Wouldn't Build Aligned AI Even If We Could · 2024-11-19T05:43:42.495Z · LW · GW

Your desire to do good and your specific proposals are valuable. But you seem to be a bit naive about power, human nature, and the difficulty of doing good even if you have power

For example, you talk about freeing people under oppressive regimes. But every extant political system and major ideology, has some corresponding notion of the greater good, and what you are calling oppressive is supposed to protect that greater good, or to protect the system against encroaching rival systems with different values. 

You mention China as oppressive and say Chinese citizens "can do [nothing] to cause meaningful improvement from my perspective". So what is it when Chinese bring sanitation or electricity to a village, or when someone in the big cities invents a new technology or launches a new service? That's Chinese people making life better for Chinese. Evidently your focus is on the one-party politics and the vulnerability of the individual to the all-seeing state. But even those have their rationales. The Leninist political system is meant to keep power in the hands of the representatives of the peasants and the workers. And the all-seeing state is just doing what you want your aligned superintelligence to do - using every means it has, to bring about the better world. 

Similar defenses can be made of every western ideology, whether conservative or liberal, progressive or libertarian or reactionary. They all have a concept of the greater good, and they all sacrifice something for the sake of it. In every case, such an ideology may also empower individuals, or specific cliques and classes, to pursue their self-interest under the cover of the ideology. But all the world's big regimes have some kind of democratic morality, as well as a persistent power elite. 

Regarding a focus on suffering - the easiest way to abolish suffering is to abolish life. All the difficulties arise when you want everyone to have life, and freedom too, but without suffering. Your principles aren't blind to this, e.g. number 3 ("spread empathy") might be considered a way to preserve freedom while reducing the possibility of cruelty. But consider number 4, "respect diversity". This can clash with your moral urgency. Give people freedom, and they may focus on their personal flourishing, rather than the suffering or oppressed somewhere else. Do you leave them to do their thing, so that the part of life's diversity which they embody can flourish, or do you lean on them to take part in some larger movement? 

I note that @daijin has already provided a different set of values which are rivals to your own. Perhaps someone could write the story of a transhuman world in which all the old politics has been abolished, and instead there's a cold war between blocs that have embraced these two value systems! 

The flip side of these complaints of mine, is that it's also not a foregone conclusion that if some group manages to create superintelligence and actually knows what they're doing - i.e. they can choose its values with confidence that those values will be maintained - that we'll just have perpetual oppression worse than death. As I have argued, every serious political ideology has some notion of the greater good, that is part of the ruling elite's culture. That elite may contain a mix of cynics, the morally exhausted and self-interested, the genuinely depraved, and those born to power, but it will also contain people who are fighting for an ideal, and new arrivals with bold ideas and a desire for change; and also those who genuinely see themselves as lovers of their country or their people or humanity, but who also have an enormously high opinion of themselves. The dream of the last kind of person is not some grim hellscape, it's a utopia of genuine happiness where they are also worshipped as transhumanity's greatest benefactor. 

Another aspect of what I'm saying, is that you feel this pessimistic about the world, because you are alienated from all the factions who actually wield power. If you were part of one of those elite clubs that actually has a chance of winning the race to create superintelligence, you might have a more benign view of the prospect that they end up wielding supreme power. 

Comment by Mitchell_Porter on Why is Gemini telling the user to die? · 2024-11-19T02:25:33.597Z · LW · GW

I don't have a detailed explanation, but the user is posting a series of assignment or exam questions. Some of them are about "abuse". Gemini is providing an example of verbal abuse. 

Comment by Mitchell_Porter on We can survive · 2024-11-05T03:09:18.297Z · LW · GW

If I understand you correctly, you want to create an unprecedentedly efficient and coordinated network, made out of intelligent people with goodwill, that will solve humanity's problems in theory and in practice? 

Comment by Mitchell_Porter on A path to human autonomy · 2024-10-29T10:51:51.598Z · LW · GW

These are my thoughts in response. I don't claim to know that what I say here is the truth, but it's a paradigm that makes sense to me. 

Strategic global cooperation to stop AI is effectively impossible, and hoping to do it by turning all the world powers into western-style democracies first is really impossible. Any successful diplomacy will have to work with the existing realities of power within and among countries, but even then, I only see tactical successes at best. Even stopping AI within the West looks very unlikely. Nationalization is conceivable, but I think it would have to partly come as an initiative from a cartel of leading companies; there is neither the will nor the understanding in the non-tech world of politics to simply impose nationalization of AI on big tech. 

For these reasons, I think the only hope of arriving at a human-friendly future by design rather than by accident, is to solve the scientific, philosophical, and design issues involved, in the creation of benevolent superhuman AI. Your idea to focus on the creation of "digital people" has a lot in common with this; more precisely, I would say that many of the questions that would have to be answered, in order to know what you're doing when creating digital people, are also questions that have to be answered, in order to know how to create benevolent superhuman AI. 

Still, in the end I expect that the pursuit of AI leads to superintelligence, and an adequately benevolent superintelligence would not necessarily be a person. It would, however, need to know what a person is, in a way that isn't tied to humanity or even to biology, because it would be governing a world in which that "unprecedented diversity of minds" can exist. 

Eliezer has argued that it is unrealistic to think that all the scientific, philosophical, and design issues can be solved in time. He also argues that in the absence of a truly effective global pause or ban, the almost inevitable upshot is a superintelligence that reorganizes the world in a way that is unfriendly to human beings, because human values are complex, and so human-friendliness requires a highly specific formulation of what human values are, and of the AI architecture that will preserve and extrapolate them. 

The argument that the design issues can't be resolved in time is strong. They involve a mix of perennial philosophical questions like the nature of the good, scientific questions like the nature of human cognition and consciousness, and avantgarde computer-science issues like the dynamics of superintelligent deep learning systems. One might reasonably expect it to take decades to resolve all these. 

Perhaps the best reason for hope here, is the use of AI as a partner in solving the problems. Of course this is a common idea, e.g. "weak-to-strong generalization" would be a form of this. It is at least conceivable that the acceleration of discovery made possible by AI, could be used to solve all the issues pertaining to friendly superintelligence, in years or months, rather than requiring decades. But there is also a significant risk that some AI-empowered group will be getting things wrong, while thinking that they are getting it right. It is also likely that even if a way is found to walk the path to a successful outcome (however narrow that path may be), that all the way to the end, there will be rival factions who have different beliefs about what the correct path is. 

As for the second proposition I have attributed to Eliezer - that if we don't know what we're doing when we cross the threshold to superintelligence, doom is almost inevitable - that's less clear to me. Perhaps there are a few rough principles which, if followed, greatly increase the odds in favor of a world that has a human-friendly niche somewhere in it. 

Comment by Mitchell_Porter on The Personal Implications of AGI Realism · 2024-10-22T17:36:06.362Z · LW · GW

Who said biological immortality (do you mean a complete cure for ageing?) requires nanobots?

We know individual cell lines can go on indefinitely, the challenge is to have an intelligent multicellular organism that can too.

Comment by Mitchell_Porter on How I'd like alignment to get done (as of 2024-10-18) · 2024-10-19T08:08:01.945Z · LW · GW

It's the best plan I've seen in a while (not perfect, but has many good parts). The superalignment team at Anthropic should probably hire you.

Comment by Mitchell_Porter on The Mysterious Trump Buyers on Polymarket · 2024-10-18T17:36:24.140Z · LW · GW

Isn't this just someone rich, spending money to make it look like the market thinks Trump will win?

Comment by Mitchell_Porter on Most arguments for AI Doom are either bad or weak · 2024-10-12T15:53:43.671Z · LW · GW

Doom aside, do you expect AI to be smarter than humans? If so, do you nonetheless expect humans to still control the world? 

Comment by Mitchell_Porter on My motivation and theory of change for working in AI healthtech · 2024-10-12T01:53:13.759Z · LW · GW

"Successionism" is a valuable new word.

Comment by Mitchell_Porter on Embracing complexity when developing and evaluating  AI responsibly · 2024-10-11T20:54:59.724Z · LW · GW

My apologies. I'm usually right when I guess that a post has been authored by AI, but it appears you really are a native speaker of one of the academic idioms that AIs have also mastered. 

As for the essay itself, it involves an aspect of AI safety or AI policy that I have neglected, namely, the management of socially embedded AI systems. I have personally neglected this in favor of SF-flavored topics like "superalignment" because I regard the era in which AIs and humans have a coexistence in which humans still have the upper hand as a very temporary thing. Nonetheless, we are still in that era right now, and hopefully some of the people working within that frame, will read your essay and comment. I do agree that the public health paradigm seems like a reasonable source of ideas, for the reasons that you give. 

Comment by Mitchell_Porter on AI: The Philosopher's Stone of the 21st Century · 2024-10-11T19:56:32.873Z · LW · GW

There's a lot going on in this essay, but the big point would appear to be: to create advanced AI is to materialize an Unknown Unknown, and why on earth would you expect that to be something you can even understand, let alone something that is sympathetic to you or "aligned" with you? 

Then I made a PDF of the article and fed it to Claude Opus and to Google's Gemini-powered NotebookLM, and both AIs seemed to get the gist immediately, as well as understanding the article's detailed structure. There is a deep irony in hearing NotebookLM's pod-people restating the essay's points in their own words, and agreeing that its warnings make sense. 

Comment by Mitchell_Porter on Embracing complexity when developing and evaluating  AI responsibly · 2024-10-10T22:55:44.530Z · LW · GW

(edit: looks like I spoke too soon and this essay is 100% pure, old-fashioned, home-grown human)

This appears to be yet another post that was mostly written by AI. Such posts are mostly ignored. 

This may be an example of someone who is not a native English speaker, using AI to make their prose more idiomatic. But then we can't tell how many of the ideas come from the AI as well. 

If we are going to have such posts, it might be useful to have them contain an introductory note, that says something about the process whereby they were generated, e.g. "I wrote an outline in my native language, and then [specific AI] completed the essay in English", or, "This essay was generated by the following prompt... this was the best of ten attempts", and so on. 

Comment by Mitchell_Porter on sarahconstantin's Shortform · 2024-10-10T20:44:13.029Z · LW · GW

Is it too much to declare this the manifesto of a new philosophical school, Constantinism?

Comment by Mitchell_Porter on why won't this alignment plan work? · 2024-10-10T18:14:55.851Z · LW · GW

each one is annotated with how much utility we estimate it to have

How are these estimates obtained?

Comment by Mitchell_Porter on If I have some money, whom should I donate it to in order to reduce expected P(doom) the most? · 2024-10-07T06:58:47.088Z · LW · GW

Let me first say what I think alignment (or "superalignment") actually requires. This is under the assumption that humanity's AI adventure issues in a superintelligence that dominates everything, and that the problem to be solved is how to make such an entity compatible with human existence and transhuman flourishing. If you think the future will always be a plurality of posthuman entities, including enhanced former humans, with none ever gaining an irrevocable upper hand (e.g. this seems to be one version of e/acc); or if you think the whole race towards AI is folly and needs to be stopped entirely; then you may have a different view. 

I have long thought of a benevolent superintelligence as requiring three things: superintelligent problem-solving ability; the correct "value system" (or "decision procedure", etc); and a correct ontology (and/or the ability to improve its ontology). The first two criteria would not be surprising, in the small world of AI safety that existed before the deep learning revolution. They fit a classic agent paradigm like the expected utility maximizer; alignment (or Friendliness, as we used to say), being a matter of identifying the right utility function. 

The third criterion is a little unconventional, and my main motive for it even more so, in that I don't believe the theories of consciousness and identity that would reduce everything to "computation". I think they (consciousness and identity) are grounded in "Being" or "substance", in a way that the virtual state machines of computation are not; that there really is a difference between a mind and a simulation of a mind, for example. This inclines me to think that quantum holism is part of the physics of mind, but that thinking of it just as physics is not enough, you need a richer ontology of which physics is only a formal description; but these are more like the best ideas I've had, than something I am absolutely sure is true. I am much more confident that purely computational theories of consciousness are radically incomplete, than as to what the correct alternative paradigm is. 

The debate about whether the fashionable reductionist theory of the day is correct, is as old as science. What does AI add to the mix? On the one hand, there is the possibility that an AI with the "right" value system but the wrong ontology, might do something intended as benevolent, that misses the mark because it misidentifies something about personhood. (A simple example of this might be, that it "uploads" everyone to a better existence, but uploads aren't actually conscious, they are just simulations.) On the other hand, one might also doubt the AI's ability to discover that the ontology of mind, according to which uploads are conscious, is wrong, especially if the AI itself isn't conscious. If it is superintelligent, it may be able to discover a mismatch between standard human concepts of mind, extrapolated in a standard way, and how reality actually works; but lacking consciousness itself, it might also lack some essential inner guidance on how the mismatch is to be corrected. 

This is just one possible story about what we could call a philosophical error in the AI's cognition and/or the design process that produced it. I think it's an example of why Wei Dai regards metaphilosophy as an important issue for alignment. Metaphilosophy is the (mostly philosophical) study of philosophy, and includes questions like, what is philosophical thought, what characterizes correct philosophical thought, and, how do you implement correct philosophical thought in an AI? Metaphilosophical concerns go beyond my third criterion, of getting ontology of mind correct; philosophy could also have something to say about problem-solving and about correct values, and even about the entire three-part approach to alignment with which I began. 

So perhaps I will revise my superalignment schema and say: a successful plan for superalignment needs to produce problem-solving superintelligence (since the superaligned AI is useless if it gets trampled by a smarter unaligned AI), a sufficiently correct "value system" (or decision procedure or utility function), and some model of metaphilosophical cognition (with particular attention to ontology of mind). 

Comment by Mitchell_Porter on Whimsical Thoughts on an AI Notepad: Exploring Non-Invasive Neural Integration via Viral and Stem Cell Pathways · 2024-10-06T21:59:24.630Z · LW · GW

Your title begins "Whimsical Thoughts on an AI Notepad", so I presume this was written with AI assistance. Please say something about your methods. Did you just supply a prompt? If so, what was it? If you did contribute more than just a prompt, what was your contribution? 

Comment by Mitchell_Porter on You can, in fact, bamboozle an unaligned AI into sparing your life · 2024-10-03T01:19:17.190Z · LW · GW

I think the common sense view is that this similarity of decision procedures provides exactly zero reason to credit the child with the fireman's decisions. Credit for a decision goes to the agent who makes it, or perhaps to the algorithm that the agent used, but not to other agents running the same or similar algorithms. 

Comment by Mitchell_Porter on Nathan Helm-Burger's Shortform · 2024-10-02T01:44:30.384Z · LW · GW

A few comments: 

The name of the paper is currently missing... 

Penrose may be a coauthor, but the estimate would really be due to his colleague Stuart Hameroff (who I worked for once), an anesthesiologist who has championed the idea of microtubules as a locus of cognition and consciousness. 

As I said in the other thread, even if microtubules do have a quantum computational role, my own expectation is that they would each only contribute a few "logical qubits" at best, e.g. via phase factors created by quasiparticles propagating around the microtubule cylinder, something which might be topologically protected against room-temperature thermal fluctuations. 

But there are many ideas about their dynamics, so, I am not dogmatic about this scenario. 

Seems like if the scientific community suspected this were true, then at least a few scientists would be trying to develop such a BCI system?

There's no mystery as to whether the scientific community suspects that the Hameroff-Penrose theory is correct. It is a well-known hypothesis in consciousness studies, but it wouldn't have too many avowed adherents beyond its inventors: perhaps some people whose own research program overlaps with it, and some other scattered sympathizers. 

It could be compared to the idea that memories are stored in RNA, another hypothesis that has been around for decades, and which has a pop-culture charisma far beyond its scientific acceptance. 

So it's not a mainstream hypothesis, but it is known enough, and overlaps with a broader world of people working on microtubule physics, biocomputation, quantum biology, and other topics. See what Google Scholar returns for "microtubule biocomputer" and scroll through a few pages. You will find, for example, a group in Germany trying to create artificial microtubule lattices for "network-based biocomputation" (which is about swarms of kinesins walking around the lattice, like mice in a maze looking for the exit), a book on using actin filaments (a relative of the microtubule) for "revolutionary computing systems", and many other varied proposals. 

I don't see anyone specifically trying to hack mouse neurons in order to make novel microtubular deep-learning systems, but that just means you're going to be the godfather of that particular concept (like Petr Vopěnka, whose "principle" started as a joke he didn't believe). 

Comment by Mitchell_Porter on You can, in fact, bamboozle an unaligned AI into sparing your life · 2024-10-01T03:25:37.133Z · LW · GW

when a child sees a fireman pull a woman out of a burning building and says "if I were that big and strong, I would also pull people out of burning buildings", in a sense it's partially the child's decision that does the work of saving the woman... but vanishingly little of the credit goes to the child

The child is partly responsible - to a very small but nonzero degree - for the fireman's actions, because the child's personal decision procedure has some similarity to the fireman's decision procedure? 

Is this a correct reading of what you said? 

Comment by Mitchell_Porter on Open Thread Summer 2024 · 2024-09-30T20:39:02.897Z · LW · GW

Most reasoning about many worlds, by physicist fans of the interpretation, as well as by non-physicists, is done in a dismayingly vague way. If you want a many-worlds framework that meets physics standards of actual rigor, I recommend thinking in terms of the consistent or decoherent histories of Gell-Mann and Hartle (e.g.). 

In ordinary quantum mechanics, to go from the wavefunction to reality, you first specify which "observable" (potentially real property) you're interested in, and then in which possible values of that observable. E.g. the observable could be position and the values could be specific possible locations. In a "Hartle multiverse", you think in terms of the history of the world, then specific observables at various times (or times + locations) in that history, then sets of possible values of those observables. You thereby get an ensemble of possible histories - all possible combinations of the possible values. The calculational side of the interpretation then gives you a probability for each possible history, given a particular wavefunction of the universe. 

For physicists, the main selling point of this framework is that it allows you to do quantum cosmology, where you can't separate the observer from the physical system under investigation. For me, it also has the advantage of being potentially relativistic, a chronic problem of less sophisticated approaches to many worlds, since spatially localized observables can be ordered in space-time rather than requiring an artificial universal time. 

On the other hand, this framework doesn't tell you how many "worlds" there are. That depends on the choice of observables. You can pick a single observable from one moment in the history of the universe (e.g. electromagnetic field strength at a certain space-time location), and use only that to define your possible worlds. That's OK if you're only interested in calculation, but if you're interested in ontology as well (also known as "what's actually there"), you may prefer some kind of "maximally refined" or "maximally fine-grained" set of histories, in which the possible worlds are defined by a set of observables and counterfactual properties that are as dense as possible while still being decoherent (e.g. without crowding so close as to violate the uncertainty principle). Investigation of maximally refined, decoherent multiverses could potentially lead to a new kind of ontological interpretation, but the topic is little investigated. 

Comment by Mitchell_Porter on You can, in fact, bamboozle an unaligned AI into sparing your life · 2024-09-30T09:17:29.350Z · LW · GW

My default hypothesis is that AI won't be even bothered by all the simulation arguments that are mindboggling to us.

I have similar thoughts, though perhaps for a different reason. There are all these ideas about acausal trade, acausal blackmail, multiverse superintelligences shaping the "universal prior", and so on, which have a lot of currency here. They have some speculative value; they would have even more value as reminders of the unknown, and the conceptual novelties that might be part of a transhuman intelligence's worldview; but instead they are elaborated in greatly varied (and yet, IMO, ill-founded) ways, by people for whom this is the way to think about superintelligence and the larger reality.

It reminds me of the pre-2012 situation in particle physics, in which it was correctly anticipated that the Higgs boson exists, but was also incorrectly expected that it would be accompanied by other new particles and a new symmetry, involved in stabilizing its mass. Thousands, maybe tens of thousands of papers were produced, proposing specific detectable new symmetries and particles that could provide this mechanism. Instead only the Higgs has shown up, and people are mostly in search of a different mechanism.

The analogy for AI would be: important but more straightforward topics have been neglected in favor of these fashionable possibilities, and, when reality does reveal a genuinely new aspect, it may be something quite different to what is being anticipated here.

Comment by Mitchell_Porter on Alexander Gietelink Oldenziel's Shortform · 2024-09-30T04:50:28.952Z · LW · GW

Thomas Kwa just provided a good reason: "measure drops off exponentially with program length". So embeddings of programs within other programs - which seems to be what a simulation is, in the Solomonoff framework - are considered exponentially unlikely. 

edit: One could counterargue that programs simulating programs increase exponentially in number. Either way, I want to see actual arguments or calculations.  

Comment by Mitchell_Porter on Alexander Gietelink Oldenziel's Shortform · 2024-09-30T00:48:39.816Z · LW · GW

I was expecting an argument like "most of the probability measure for a given program, is found in certain embeddings of that program in larger programs". Has anyone bothered to make a quantitative argument, a theorem, or a rigorous conjecture which encapsulates this claim?

Comment by Mitchell_Porter on Alexander Gietelink Oldenziel's Shortform · 2024-09-29T16:47:28.119Z · LW · GW

Please demonstrate that the Solomonoff prior favors simulation.

Comment by Mitchell_Porter on The Compute Conundrum: AI Governance in a Shifting Geopolitical Era · 2024-09-29T02:12:47.148Z · LW · GW

Please describe the prompts and general process you used to create this essay...

Comment by Mitchell_Porter on Non-human centric view of existence · 2024-09-25T03:44:21.425Z · LW · GW

I am in bad shape but I will try to answer. I was interested in responding to your question but wanted more context. And it seems from what you say, that the issue is not at all about whether humanity or individual humans live forever, but whether one should be attached to their existence, or whether the pursuit of some EA-style greater good would always favor their existence. 

I mean, your friend seems to be suggesting a blase attitude towards the prospect of human extinction via AI. Such an attitude has some precedent in the stoic attitude towards individual death. One aims to be unaffected by it, by telling oneself that this is the natural order. With humanity menaced by extinction that one may be helpless to prevent, an individual might analogously find some peace by telling themselves it's all just nature. 

But someone once observed that people who aim to be philosophically accepting about the death of humanity, would often be far more concerned about the death of individuals, for example children. If this is the case, it suggests that they haven't really grasped what human extinction means. 

There is much more that can be said about all these issues, but that's the best that I can do for now. 

Comment by Mitchell_Porter on Non-human centric view of existence · 2024-09-24T01:22:26.373Z · LW · GW

I think I'm still missing the connection. How did you get from "AI versus humanity" to "whether human immortality is desirable"?

Comment by Mitchell_Porter on Non-human centric view of existence · 2024-09-23T22:50:25.404Z · LW · GW

why must human/humanity live/continue forever?

The status quo is that they don't. All beings, families, nations, species, worlds perish - that is what science, appearance, and experience tell us. Eventually, they always perish, and often enough they are cut off early, in a way that aborts their basic potential. 

How did this question arise?