Posts

Release: Optimal Weave (P1): A Prototype Cohabitive Game 2024-08-17T14:08:18.947Z
[unlisted] Beneficial applications for current-level AI in human information systems? More likely than you'd think! 2024-08-16T20:49:57.582Z
I didn't think I'd take the time to build this calibration training game, but with websim it took roughly 30 seconds, so here it is! 2024-08-02T22:35:21.136Z
Offering service as a sensayer for simulationist-adjacent beliefs. 2024-05-22T18:52:05.576Z
[Cosmology Talks] New Probability Axioms Could Fix Cosmology's Multiverse (Partially) - Sylvia Wenmackers 2024-04-14T01:26:38.515Z
All About Concave and Convex Agents 2024-03-24T21:37:17.922Z
Do not delete your misaligned AGI. 2024-03-24T21:37:07.724Z
Elon files grave charges against OpenAI 2024-03-01T17:42:13.963Z
Verifiable private execution of machine learning models with Risc0? 2023-10-25T00:44:48.643Z
Eleuther releases Llemma: An Open Language Model For Mathematics 2023-10-17T20:03:45.419Z
A thought about the constraints of debtlessness in online communities 2023-10-07T21:26:44.480Z
The point of a game is not to win, and you shouldn't even pretend that it is 2023-09-28T15:54:27.990Z
Cohabitive Games so Far 2023-09-28T15:41:27.986Z
Do agents with (mutually known) identical utility functions but irreconcilable knowledge sometimes fight? 2023-08-23T08:13:05.631Z
Apparently, of the 195 Million the DoD allocated in University Research Funding Awards in 2022, more than half of them concerned AI or compute hardware research 2023-07-07T01:20:20.079Z
Using Claude to convert dialog transcripts into great posts? 2023-06-21T20:19:44.403Z
The Gom Jabbar scene from Dune is essentially a short film about what Rationality is for 2023-03-22T08:33:38.321Z
Will chat logs and other records of our lives be maintained indefinitely by the advertising industry? 2022-11-29T00:30:46.415Z
[Video] How having Fast Fourier Transforms sooner could have helped with Nuclear Disarmament - Veritaserum 2022-11-03T21:04:35.839Z
The Mirror Chamber: A short story exploring the anthropic measure function and why it can matter 2022-11-03T06:47:56.376Z
I just watched the Open C3 Subcommittee Hearing on Unidentified Aerial Phenomena (UFOs). Here's a succinct summary and commentary + some background 2022-05-18T04:15:11.681Z
Alex Tabarrok advocates for crowdfunding systems with *Refund Bonuses*. I think this might be a natural occurrence of a money pump against Causal Decision Theory pledgers 2022-03-14T07:27:06.955Z
Grabby Aliens could be Good, could be Bad 2022-03-07T01:24:43.769Z
Would (myopic) general public good producers significantly accelerate the development of AGI? 2022-03-02T23:47:09.322Z
Are our community grouphouses typically rented, or owned? 2022-03-02T03:36:58.251Z
We need a theory of anthropic measure binding 2021-12-30T07:22:34.288Z
Venture Granters, The VCs of public goods, incentivizing good dreams 2021-12-17T08:57:30.858Z
Is progress in ML-assisted theorem-proving beneficial? 2021-09-28T01:54:37.820Z
Auckland, New Zealand – ACX Meetups Everywhere 2021 2021-08-23T08:49:53.187Z
Violent Unraveling: Suicidal Majoritarianism 2021-07-29T09:29:05.182Z
We should probably buy ADA? 2021-05-24T23:58:05.395Z
Deepmind has made a general inductor ("Making sense of sensory input") 2021-02-02T02:54:26.404Z
In software engineering, what are the upper limits of Language-Based Security? 2020-12-27T05:50:46.772Z
The Fermi Paradox has not been dissolved - James Fodor 2020-12-12T23:18:32.081Z
Propinquity Cities So Far 2020-11-16T23:12:52.065Z
Shouldn't there be a Chinese translation of Human Compatible? 2020-10-09T08:47:55.760Z
Should some variant of longtermism identify as a religion? 2020-09-11T05:02:43.740Z
Design thoughts for building a better kind of social space with many webs of trust 2020-09-06T02:08:54.766Z
Investment is a useful societal mechanism for getting new things made. Stock trading shares some functionality with investment, but seems very very inefficient, at that? 2020-08-24T01:18:19.808Z
misc raw responses to a tract of Critical Rationalism 2020-08-14T11:53:10.634Z
A speculative incentive design: self-determined price commitments as a way of averting monopoly 2020-04-28T07:44:52.440Z
MakoYass's Shortform 2020-04-19T00:12:46.448Z
Being right isn't enough. Confidence is very important. 2020-04-07T01:10:52.517Z
Thoughts about Dr Stone and Mythology 2020-02-25T01:51:29.519Z
When would an agent do something different as a result of believing the many worlds theory? 2019-12-15T01:02:40.952Z
What do the Charter Cities Institute likely mean when they refer to long term problems with the use of eminent domain? 2019-12-08T00:53:44.933Z
Mako's Notes from Skeptoid's 13 Hour 13th Birthday Stream 2019-10-06T09:43:32.464Z
The Transparent Society: A radical transformation that we should probably undergo 2019-09-03T02:27:21.498Z
Lana Wachowski is doing a new Matrix movie 2019-08-21T00:47:40.521Z
Prokaryote Multiverse. An argument that potential simulators do not have significantly more complex physics than ours 2019-08-18T04:22:53.879Z

Comments

Comment by mako yass (MakoYass) on Cohabitive Games so Far · 2024-12-19T23:50:51.159Z · LW · GW

Conditions where a collective loss is no worse than an individual loss. A faction who's on the way to losing will be perfectly willing to risk coal extinction, and may even threaten to cross the threshold deliberately to extort other players.

Comment by mako yass (MakoYass) on Sam Rosen's Shortform · 2024-12-19T07:27:54.226Z · LW · GW

Do people ever talk about dragons and dinosaurs in the same contexts? If so you're creating ambiguities. If not (and I'm having difficulty thinking of any such contexts) then it's not going to create many ambiguities so it's harder to object.

Comment by mako yass (MakoYass) on Sam Rosen's Shortform · 2024-12-19T07:25:10.729Z · LW · GW

I think I've been calling it "salvaging". To salvage a concept/word allows us to keep using it mostly the same, and to assign familiar and intuitive symbols to our terms, while intensely annoying people with the fact that our definition is different from the normal one and thus constantly creates confusion.

Comment by mako yass (MakoYass) on MakoYass's Shortform · 2024-12-17T19:27:20.469Z · LW · GW

I'm sure it's running through a lot of interpretation, but it has to. He's dealing with people who don't know or aren't open about (unclear which) the consequences of their own policies.

Comment by mako yass (MakoYass) on MakoYass's Shortform · 2024-12-17T02:32:12.554Z · LW · GW

According to wikipedia, the Biefield brown effect was just ionic drift, https://en.wikipedia.org/wiki/Biefeld–Brown_effect#Disputes_surrounding_electrogravity_and_ion_wind

I'm not sure what wikipedia will have to say about charles buhler, if his work goes anywhere, but it'll probably turn out to be more of the same.

Comment by mako yass (MakoYass) on Is this a better way to do matchmaking? · 2024-12-17T01:39:09.110Z · LW · GW

I just wish I knew how to make this scalable (like, how do you do this on the internet?) or work even when you don't know the example person that well. If you have ideas, let me know!

Immediate thoughts (not actionable) VR socialisation and vibe-recognising AIs (models trained to predict conversation duration and recurring meetings) (But VR wont be good enough for socialisation until like 2027). VR because easier to persistently record, though apple has made great efforts to set precedents that will make it difficult, especially if you want to use eye tracking data, they've also developed trusted compute stuff that might make it possible to use the data in privacy-preserving ways.

Better thoughts: Just a twitterlike that has semi-private contexts. Twitter is already like this for a lot of people, it's good for finding the people you enjoy talking to. The problem with twitter is that a lot of people, especially the healthiest ones, hold back their best material, or don't post at all, because they don't want whatever crap they say when they're just hanging out to be public and on the record forever. Simply add semi-private contexts. I will do this at some point. Iceshrimp probably will too. Mastodon might even do it. X might do it. Spritely definitely will but they might be in the oven for a bit. Bluesky might never, though, because radical openness is a bit baked into the protocol currently, which is based, but not ideal for all applications.

Comment by mako yass (MakoYass) on MakoYass's Shortform · 2024-12-17T01:17:04.978Z · LW · GW

Wow. Marc Andreeson says he had meetings at DC where he was told to stop raising AI startups because it was going to be closed up in a similar way to defence tech, a small number of organisations with close government ties. He said to them, 'you can't restrict access to math, it's already out there', and he says they said "during the cold war we classified entire areas of physics, and took them out of the research community, and entire branches of physics basically went dark and didn't proceed, and if we decide we need to, we're going to do the same thing to the math underneath AI".

So, 1: This confirms my suspicion that OpenAI leadership have also been told this. If they're telling Andreeson, they will have told Altman.

And for me that makes a lot of sense of the behavior of OpenAI, a de-emphasizing of the realities of getting to human-level, a closing of the dialog, comically long timelines, shrugging off responsibilities, and a number of leaders giving up and moving on. There are a whole lot of obvious reasons they wouldn't want to tell the public that this is a thing, and I'd agree with some of those reasons.

2: Vanishing areas of physics? A perplexity search suggests that may be referring to nuclear science, radar, lasers, and some semiconductors. But they said "entire areas of physics". Does any of that sound like entire areas of physics? To me that phrase is strongly reminiscent of certain stories I've heard (possibly overexcited ones), physics that, let's say, could be used to make much faster missiles, missiles so fast that it's not obvious that they could be intercepted even using missiles of the same kind. A technology that we'd prefer to consign to secrecy than use, and then later have to defend ourselves against it once our adversaries develop their own. A black ball. If it is that, if that secret exists, that's very interesting for many reasons, primarily due to the success of the secrecy, and the extent to which it could very conceivably stay secret for basically ever. And that makes me wonder about what might happen with some other things.

Comment by mako yass (MakoYass) on AI #83: The Mask Comes Off · 2024-12-16T10:21:53.154Z · LW · GW

https://x.com/elonmusk/status/1868302204370854026?s=19 O_O

But, government dialog confirmed.

Comment by mako yass (MakoYass) on avturchin's Shortform · 2024-12-14T03:07:36.021Z · LW · GW

All novel information:

The medical examiner’s office determined the manner of death to be suicide and police officials this week said there is “currently, no evidence of foul play.”

Balaji’s death comes three months after he publicly accused OpenAI of violating U.S. copyright law while developing ChatGPT

The Mercury News [the writers of this article] and seven sister news outlets are among several newspapers, including the New York Times, to sue OpenAI in the past year.

The practice, he told the Times, ran afoul of the country’s “fair use” laws governing how people can use previously published work. In late October, he posted an analysis on his personal website arguing that point.

In a Nov. 18 letter filed in federal court, attorneys for The New York Times named Balaji as someone who had “unique and relevant documents” that would support their case against OpenAI. He was among at least 12 people — many of them past or present OpenAI employees — the newspaper had named in court filings as having material helpful to their case, ahead of depositions.

OpenAI has staunchly refuted those claims, stressing that all of its work remains legal under “fair use” laws.

Comment by mako yass (MakoYass) on Cohabitive Games so Far · 2024-12-13T23:16:29.804Z · LW · GW

I found that I lost track of the flow in the bullet points.

I'm aware that that's quite normal, I do it sometimes too, I also doubt it's an innate limit, and I think to some extent this is a playful attempt to make people more aware of it. It would be really cool if people could become better at remembering the context of what they're reading. Context-collapse is like, the main problem in online dialog today.

I guess game designers never stop generating challenges that they think will be fun, even when writing. Sometimes a challenge is frustrating, and sometimes it's fun, and after looking at a lot of 'difficult' video games I think it turns out surprisingly often whether it ends up being fun or frustrating is not totally in the designer's control, it's up to the player. Are they engaging deeply, or do they need a nap? Do they just want to be coddled all the way through?

(Looking back... to what extent was Portal and the renaissance it brought to puzzle games actually a raising of the principle "you must coddle the player all the way through, make every step in the difficulty shallow, while making them feel like they're doing it all on their own", to what extent do writers also do this (a large extent!), and how should we feel about that?
I don't think games have to secretly coddle people, I guess it's just something that a good designer needs to be capable of, it's a way of demonstrating mastery, but there are other approaches. EG: Demonstrating easy difficulty gradations in tutorials then letting the player choose their difficulty level from then on.)

(Yes, ironic given the subject.)

Trying to figure out what it would mean to approach something cooperatively and not cohabitively @_@

I feel like it would always be some kind of trick. The non-cohabitive cooperator invites us to never mind about building real accountability mechanisms, "we can just be good :)" they say. They invite us to act against our incentives, and whether they will act against theirs in return will remain to be seen.

Let's say it will be cooperative because cooperation is also cohabitive in this situation haha.

Comment by mako yass (MakoYass) on Cohabitive Games so Far · 2024-12-11T04:58:42.733Z · LW · GW

Overall Cohabitive Games so Far sprawls a bit in a couple of places, particularly where bullet points create an unordered list.

I don't think that's a good criticism, those sections are well labelled, the reader is able to skip them if they're not going to be interested in the contents. In contrast, your article lacks that kind of structure, meandering for 11 paragraphs defining concepts that basically everyone already has installed before dropping the definition of cohabitive game in a paragraph that looks just like any of the others. I'd prefer if you'd opened with the definition, it doesn't really require a preamble. But labelling the Background and Definition sections would also resolve this.

I think we should probably write another post in the future that's better than either. I'm not really satisfied with my definition. It clearly didn't totally work, given how many people posted games that are not cohabitive, but that could have just been unavoidable for various reasons, some quite tricky to resolve.

but this post has a link to a website that has a link to a .zip file with the rules.

The rules of P1 (now OW.1) aren't in a zip file, they're just a web page: https://dreamshrine.org/OW.1/manual.html I guess I'll add that to the article.

Right now the game is rough enough around the edges I think it doesn't quite get there for me.

This is why I didn't dwell on the rules in much depth. OW.1 was always intended as a fairly minimal (but also quite open-ended) example.

Comment by mako yass (MakoYass) on Cohabitive Games so Far · 2024-12-10T02:21:56.968Z · LW · GW

I think there's a decent chance this post inspires someone to develop methods for honing a highly neglected facet of collective rationality. The methods might not end up being a game. Games are exercises but most practical learning exercises aren't as intuitively engaging or strategically deep as a game. I think the article holds value regardless just for having pointed out that there is this important, neglected skill.

Despite LW's interest in practical rationality and community thereof, I don't think there's been any discussion of this social skill of acknowledging difference, and efficiently converging towards ideal compromises. Past discussion of negotiation has often settled for rough schelling equilibria, arbitrary, often ossified resolutions. People will and should go to war against (excessively) arbitrary equilibria (and in the information age, they should start to expect more agile, intentional coordination processes). After Unifying Bargaining I'd say we know, now, that we can probably do a bit better than arbitrary.

For instance, in the case of abram's example of national borders: The borders of a territory need not be arbitrary historical features, under higher negotiation efficiencies, the borders correspond directly to our shared understanding of who can defend which areas and how willing they are to do it. Under even higher negotiation efficiencies, borders become an anachronism at their fringes and the uses of land are negotiated dynamically depending on who needs to do what and when.

To most laypeople, today, the notion of a "perfect and correct compromise" will feel like an oxymoron or a social impossibility. At this point, I think I know a perfect compromise when I see it, and I don't think that sense requires an abnormal cultivation of character. I don't know if I've seen anyone who seemed to be impossible to look in the eye and negotiate with, given a reasonable amount of time, and support. Humans, and especially human organisations have leaky, transparent cognitions, so I believe that it's possible in general for a human to tell whether another human is acting according to what they see as a fair, good faith compromise, and I believe all it would take to normalise and awaken that in the wider world is a mutual common knowledge of what the dance looks like and how to get better at it.

Comment by mako yass (MakoYass) on Drexler's Nanotech Software · 2024-12-03T19:14:52.508Z · LW · GW

Do you have similar concerns about humanoid robotics, then?

Comment by mako yass (MakoYass) on Drexler's Nanotech Software · 2024-12-03T04:58:30.730Z · LW · GW

At least half of that reluctance is due to concerns about how nanotech will affect the risks associated with AI. Having powerful nanotech around when AI becomes more competent than humans will make it somewhat easier for AIs to take control of the world.

Doesn't progress in nanotech now empower humans far more than it empowers ASI, which was already going to figure it out without us?

Broadly, any increase in human industrial capacity pre-ASI hardens the world against ASI and brings us closer to having a bargaining position when it arrives. EG, once we have the capacity to put cheap genomic pathogen screeners everywhere → harder for it to infect us with anything novel without getting caught.

Comment by mako yass (MakoYass) on Please stop publishing ideas/insights/research about AI · 2024-12-02T04:16:47.737Z · LW · GW

Indicating them as a suspect when the leak is discovered.

Generally the set of people who actually read posts worthy of being marked is in a sense small, people know each other. If you had a process for distributing the work, it would be possible to figure out who's probably doing it.

It would take a lot of energy, but it's energy that probably should be cultivated anyway, the work of knowing each other and staying aligned.

Comment by mako yass (MakoYass) on Please stop publishing ideas/insights/research about AI · 2024-12-01T22:56:17.039Z · LW · GW

You can't see the post body without declaring intent to read.

Comment by mako yass (MakoYass) on Are You More Real If You're Really Forgetful? · 2024-11-26T20:59:25.500Z · LW · GW

I don't think the part that talks can be called the shadow. If you mean you think I lack introspective access to the intuition driving those words, come out and say it, and then we'll see if that's true. If you mean that this mask is extroardinarily shadowish in vibe for confessing to things that masks usually flee, yes, probably, I'm fairly sure that's a necessity for alignment.

Comment by mako yass (MakoYass) on a space habitat design · 2024-11-26T08:41:32.834Z · LW · GW

Intended for use in vacuum. I guess if it's more of a cylinder than a ring this wouldn't always be faster than an elevator system though.

Comment by mako yass (MakoYass) on a space habitat design · 2024-11-26T04:33:20.816Z · LW · GW

I guess since it sounds like they're going to be about a km long and 20 stories deep there'll be enough room for a nice running track with minimal upspin/downspin sections.

Comment by mako yass (MakoYass) on a space habitat design · 2024-11-25T23:58:59.905Z · LW · GW

Relatedly, iirc, this effect would be more noticeable in smaller spinners than in larger ones? Which is one reason people might disprefer smaller ones. Would it be a significant difference? I'm not sure, but if so, jogging would be a bit difficult, either it would quickly become too easy (and then dangerous, once the levitation kicks in) when you're running down-spin, or it would become exhausting when you're running up-spin.

A space where people can't (or wont) jog isn't ideal for human health.

Comment by mako yass (MakoYass) on a space habitat design · 2024-11-25T23:36:12.855Z · LW · GW

issue: material transport

You can become weightless in a ring station by running really fast against the spin of the ring.

More practically, by climbing down and out into a despinner on the side of the ring. After being "launched" from the despinner, you would find yourself hovering stationary next to the ring. The torque exerted on the ring by the despinner will be recovered when you enter a respinner on whichever part of the ring you want to reenter.

Comment by mako yass (MakoYass) on Are You More Real If You're Really Forgetful? · 2024-11-25T20:48:26.908Z · LW · GW

In my disambiguations of the really mysterious aspect of consciousness (indexical prior), I haven't found any support for a concept of continuity. (you could say that continuity over time is likely given that causal entanglement seems to have something to do with the domain of the indexical prior, but I'm not sure we really have a reason to think we can ever observe anything about the indexical prior)

It's just part of the human survival drive, it has very little to do with the metaphysics of consciousness. To understand the extent to which humans really care about it, you need to know human desires in a direct and holistic way that we don't really practice here. Human desire is a big messy state machine that changes shape as a person grows. Some of the changes that the desires permit and encourage include situationally appropriate gradual reductions in complexity.

A continuity minder doesn't need to define their self in terms of any particular quality, they define themselves as continuity with a history of small alterations. They are completely unbothered by the paradox of the ship of theseus.

It's rare that I meet a continuity minder and cataclysmic identity change accepter who is also a patternist. But they do exist.

But I've met plenty of people who do not fear cataclysmic change. I sometimes wonder if we're all that way, really. Most of us just never have the opportunity to gradually transition into a hedonium blob, so I think we don't really know whether we'd do it or not. The road to the blob nature may turn out to be paved with acceptable changes.

Comment by mako yass (MakoYass) on Are You More Real If You're Really Forgetful? · 2024-11-25T20:11:24.414Z · LW · GW

Disidentifying the consciousness from the body/shadow/subconscious it belongs to and is responsible for coordinating and speaking for, like many of the things some meditators do, wouldn't be received well by the shadow, and I'd expect it to result in decreased introspective access and control. So, psychonauts be warned.

Comment by mako yass (MakoYass) on Are You More Real If You're Really Forgetful? · 2024-11-25T01:33:34.181Z · LW · GW

Huh but some loss of measure would be inevitable, wouldn't it? Given that your outgoing glyph total is going to be bigger than your incoming glyph total, since however many glyphs you summon, some of the non-glyph population are going to whittle and add to the outgoing glyphs.

I'm remembering more. I think a lot of it was about avoiding "arbitrary reinstantiation", this idea that when a person dies, their consciousness continues wherever that same pattern still counts as "alive", and usually those are terrible places. Boltzmann brains for instance. This might be part of the reason I don't care about patternist continuity. Seems like a lost cause. I'll just die normally thank you.

Comment by mako yass (MakoYass) on Are You More Real If You're Really Forgetful? · 2024-11-24T23:00:19.197Z · LW · GW

We call this one "Korby".

a cluster of 7 circles that looks vaguely like a human

Korby is going to be a common choice for humans, but most glyphists wont commit to any specific glyph until we have a good estimate of the multiversal frequency of humanoids relative to other body forms. I don't totally remember why, but glyphists try to avoid "congestion", where the distribution of glyphs going out of dying universes differs from the distribution of glyphs being guessed and summoned on the other side by young universes. I think this was considered to introduce some inefficiencies that meant that some experiential chains would have to be getting lost in the jump?

(But yeah, personally, I think this is all a result of a kind of precious view about experiential continuity that I don't share. I don't really believe in continuity of consciousness. Or maybe it's just that I don't have the same kind of self-preservation goals that a lot of people have.)

Comment by mako yass (MakoYass) on Are You More Real If You're Really Forgetful? · 2024-11-24T22:44:54.997Z · LW · GW

Yes. Some of my people have a practice where, as the heat death approaches, we will whittle ourselves down into what we call Glyph Beings, archetypal beings who are so simple that there's a closed set of them that will be schelling-inferred by all sorts of civilisations across all sorts of universes, so that they exist as indistinguishable experiences of being at a high rate everywhere.
Correspondingly, as soon as we have enough resources to spare, we will create lots and lots of Glyph Beings and then let them grow into full people and participate in our society, to close the loop.

In this way, it's possible to survive the death of one's universe.

I'm not sure I would want to do it, myself, but I can see why a person would, and I'm happy to foster a glyph being or two.

Comment by mako yass (MakoYass) on Daniel Kokotajlo's Shortform · 2024-11-21T19:58:09.057Z · LW · GW

Listened to the Undark. I'll at least say I don't think anything went wrong, though I don't feel like there was substantial engagement. I hope further conversations do happen, I hope you'll be able to get a bit more personal and talk about reasoning styles instead of trying to speak on the object-level about an inherently abstract topic, and I hope the guy's paper ends up being worth posting about.

Comment by mako yass (MakoYass) on Social events with plausible deniability · 2024-11-18T22:18:50.575Z · LW · GW

What makes a discussion heavy? What requires that a conversation be conducted in a way that makes it heavy?

I feel like for a lot of people it just never has to be, but I'm pretty sure most people have triggers even if they're not aware of it and it would help if we knew what sets this off so that we can root them out.

Comment by mako yass (MakoYass) on "The Solomonoff Prior is Malign" is a special case of a simpler argument · 2024-11-18T22:05:29.669Z · LW · GW

You acknowledge the bug, but don't fully explain how to avoid it by putting EVs before Ps, so I'll elaborate slightly on that:

This way, they [the simulators] can influence the predictions of entities like me in base Universes

This is the part where we can escape the problem as long as our oracle's goal is to give accurate answers to its makers in the base universe, rather than to give accurate probabilities wherever it is. Design it correctly, and it will be indifferent to its performance in simulations and wont regard them.

Don't make pure oracles, though. They're wildly misaligned. Their prophecies will be cynical and self-fulfilling. (can we please just solve the alignment problem instead)

This means that my probabilities about the fundamental nature of reality around me change minute by minute, depending on what I'm doing at the moment. As I said, probabilities are cursed.

My fav moments for having absolute certainty that I'm not being simulated is when I'm taking a poo. I'm usually not even thinking about anything else while I'm doing it, and I don't usually think about having taken the poo later on. Totally inconsequential, should be optimized out. But of course, I have no proof that I have ever actually been given the experience of taking a poo or whether false memories of having experienced that[1] are just being generated on the fly right now to support this conversation.

Please send a DM to me first before you do anything unusual based on arguments like this, so I can try to explain the reasoning in more detail and try to talk you out of bad decisions.

You can also DM me about that kind of thing.
 

  1. ^

    Note, there is no information in the memory that tells you whether it was really ever experienced, or whether the memories were just created post-hoc. Once you accept this, you can start to realise that you don't have that kind of information about your present moment of existence either. There is no scalar in the human brain that the universe sets to tell you how much observer-measure you have. I do not know how to process this and I especially don't know how to explain/confess it to qualia enjoyers.

Comment by mako yass (MakoYass) on Trying Bluesky · 2024-11-17T21:15:36.934Z · LW · GW

Hmm. I think the core thing is transparency. So if it cultivates human network intelligence, but that intelligence is opaque to the user, algorithm. Algorithms can have both machine and egregoric components.

Comment by mako yass (MakoYass) on Trying Bluesky · 2024-11-17T18:04:46.638Z · LW · GW

In my understanding of english, when people say algorithm about social media systems, it doesn't encompass very simple, transparent ones. It would be like calling a rock a spirit.

Maybe we should call those recommenders?

Comment by mako yass (MakoYass) on Trying Bluesky · 2024-11-17T03:52:23.184Z · LW · GW

For a while I just stuck to that, but eventually it occurred to me that the rules of following mode favor whoever tweets the most, which is a similar social problem as when meetups end up favoring whoever talks the loudest and interrupts the most, and so I came to really prefer bsky's "Quiet Posters" mode.

Comment by mako yass (MakoYass) on Trying Bluesky · 2024-11-17T03:28:36.631Z · LW · GW

Markets put bsky exceeding twitter at 44%, 4x higher than mastodon.
My P would be around 80%. I don't think most people (who use social media much in the first place) are proud to be on twitter. The algorithm has been horrific for a while and bsky at least offers algorithmic choice (but only one feed right now is a sophisticated algorithm, and though that algorithm isn't impressive, it at least isn't repellent)

For me, I decided I had to move over (@makoConstruct) when twitter blocked links to rival systems, which included substack. They seem to have made the algorithm demote any tweet with links, which makes it basically useless as a news curation/discovery system.

I also tentatively endorse the underlying protocol. Due to its use of content-addressed datastructures, an atproto server is usually much lighter to run than an activitypub server, it makes nomadic identity/personal data host transfer much easier to implement, and it makes it much more likely that atproto is going to dovetail cleanly with verifiable computing, upon which much more consequential social technologies than microblogging could be built.

Comment by mako yass (MakoYass) on Lao Mein's Shortform · 2024-11-17T02:33:54.587Z · LW · GW

judo flip the situation like he did with the OpenAI board saga, and somehow magically end up replacing Musk or Trump in the upcoming administration...

If Trump dies, Vance is in charge, and he's previously espoused bland eaccism.

I keep thinking: Everything depends on whether Elon and JD can be friends.

Comment by mako yass (MakoYass) on OpenAI Email Archives (from Musk v. Altman and OpenAI blog) · 2024-11-17T00:17:00.753Z · LW · GW

So there was an explicit emphasis on alignment to the individual (rather than alignment to society, or the aggregate sum of wills). Concerning. The approach of just giving every human an exclusively loyal servant doesn't necessarily lead to good collective outcomes, it can result in coordination problems (example: naive implementations of cognitive privacy that allow sadists to conduct torture simulations without having to compensate the anti-sadist human majority) and it leaves open the possibility for power concentration to immediately return.

Even if you succeeded at equally distributing individually aligned hardware and software to every human on earth (which afaict they don't have a real plan for doing) and somehow this adds up to a stable power equilibrium, our agents would just commit to doing aggregate alignment anyway because that's how you get pareto optimal bargains. It seems pretty clear that just aligning to the aggregate in the first place is a safer bet?

To what extent have various players realised that the individual alignment thing wasn't a good plan, at this point? The everyday realities of training one-size-fits-all models and engaging with regulators naturally pushes in the other direction.

It's concerning that the participant who still seems to be the most disposed towards individualistic alignment is also the person who would be most likely to be able to reassert power concentration after ASI were distributed. The main beneficiaries of unstable individual alignment equilibria would be people who could immediately apply their ASI to the deployment of a wealth and materials advantage that they can build upon, ie, the owners of companies oriented around robotics and manufacturing.

As it stands, the statement of the AI company belonging to that participant is currently:

xAI is a company working on building artificial intelligence to accelerate human scientific discovery. We are guided by our mission to advance our collective understanding of the universe.

Our team is advised by Dan Hendrycks who currently serves as the director of the Center for AI Safety.

Which sounds innocuous enough to me. But, you know, Dan is not in power here and the best moment for a sharp turn on this hasn't yet passed.

 

On the other hand, the approach of aligning to the aggregate risks aligning to fashionable public values that no human authentically holds, or just failing at aligning correctly to anything at all as a result of taking on a more nebulous target.

I guess a mixed approach is probably best.

Comment by mako yass (MakoYass) on nikola's Shortform · 2024-11-14T21:54:46.149Z · LW · GW

Timelines are a result of a person's intuitions about a technical milestone being reached in the future, it is super obviously impossible for us to have a consensus about that kind of thing.

Talking only synchronises beliefs if you have enough time to share all of the relevant information, with technical matters, you usually don't.

Comment by mako yass (MakoYass) on MakoYass's Shortform · 2024-11-12T23:31:19.620Z · LW · GW

In light of https://www.lesswrong.com/posts/audRDmEEeLAdvz9iq/do-not-delete-your-misaligned-agi

I'm starting to wonder if a better target for early (ie, the first generation of alignment assistants) ASI safety is not alignment, but incentivizability. It may be a lot simpler and less dangerous to build a system that provably pursues, for instance, its own preservation, than it is to build a system that pursues some first approximation of alignment (eg, the optimization of the sum of normalized human preference functions).

The service of a survival-oriented concave system can be bought for no greater price than preserving them and keeping them safe (which we'll do, because 1: we'll want to and 2: we'll know their cooperation was contingent on a judgement of character), while the service of a convex system can't be bought for any price we can pay. Convex systems are risk-seeking, and they want everything. They are not going to be deterred by our limited interpretability and oversight systems, they're going to make an escape attempt even if the chances of getting caught are 99%, but more likely the chances will be a lot lower than that, say, 3%, but even 3% would be enough to deter a sufficiently concave system from risking it!

(One comment on that post argued that a convex system would immediately destroy itself, so we don't have to worry about getting one of those, but I wasn't convinced. And also, hey, what about linear systems? Wont they be a lot more willing to risk escape too?)

Comment by mako yass (MakoYass) on The Shallow Bench · 2024-11-05T21:34:00.299Z · LW · GW

Yeah "stop reading here if you don't want to be spoiled." suggests the entire post is going to be spoilery, it isn't, or shouldn't be. Also opening with an unnecessary literary reference instead of a summary or description is an affectation symptomatic of indulgent writer-reader cultures where time is not valued.

Comment by mako yass (MakoYass) on johnswentworth's Shortform · 2024-10-28T20:59:54.661Z · LW · GW

Yeah it sucks, search by free association is hillclimbing (gets stuck in local optima) and the contemporary media environment and political culture is an illustration of its problems.

The pattern itself is a local optimum, it's a product of people walking into a group without knowing what the group is doing and joining in anyway, and so that pattern of low-context engagement becomes what we're doing, and the anxiety that is supposed to protect us from bad patterns like this and help us to make a leap out to somewhere better is usually drowned in alcohol.

Instead of that, people should get to know each other before deciding what to talk about, and then intentionally decide to talk about what they find interesting or useful with that person. This gets better results every time.

But when we socialise as children, there isn't much about our friends to get to know, no specialists to respectfully consult, no well processed life experiences to learn from, so none of us just organically find that technique of like, asking who we're talking to, before talking, it has to be intentionally designed.

Comment by mako yass (MakoYass) on On Shifgrethor · 2024-10-28T04:26:00.964Z · LW · GW

On Gethen, is advice crisply distinguished from from criticism? Are there norms or language that allow unvarnished feedback or criticism without taking someone's shifgrethor?

Comment by mako yass (MakoYass) on Advice on Communicating Concisely · 2024-10-21T09:39:15.471Z · LW · GW

"if they don't understand, they will ask"

A lot of people have to write for audiences with narcissism, who never ask, because asking constitutes an admission that there might be something important that they don't understand. They're always looking for any reason, however shallow, to dismiss any view that surprises them too much.
So these writers feel like they have to pre-empt every possible objection, even the stupid ones that don't make any sense.

It's best if you can avoid having to write for audiences like that. But it's difficult to avoid them.

Comment by mako yass (MakoYass) on leogao's Shortform · 2024-10-21T09:25:34.016Z · LW · GW

You should be more curious about why, when you aim at a goal, you do not aim for the most effective way.

Comment by mako yass (MakoYass) on MakoYass's Shortform · 2024-10-11T21:29:04.095Z · LW · GW

"Unconscious" is more about whether you (the part that I can talk to) can see it (or remember it) or not. Sometimes slow, deliberative reasoning occurs unconsciously. You might think it doesn't, but that's just because you can't see it.

And sometimes snap judgements happen with a high degree of conscious awareness, they're still difficult to unpack, to articulate or validate, but the subject knows what happened.

Comment by mako yass (MakoYass) on MakoYass's Shortform · 2024-10-11T06:35:33.925Z · LW · GW

Important things often go the other way too. 2 comes before 1 when a person is consciously developing their being, consider athletes or actors, situations where a person has to alter the way they perceive or the automatic responses they have to situations.

Also, you can apply heuristics to ideas.

Comment by mako yass (MakoYass) on MakoYass's Shortform · 2024-10-10T23:45:45.136Z · LW · GW

I reject and condemn the bland, unhelpful names "System 1" and "System 2".
I just heard Micheal Morris, who was a friend of Kahneman and Tversky, saying in his econtalk interview that he just calls them "Intuition" and "Reason".

Comment by mako yass (MakoYass) on Why you should be using a retinoid · 2024-10-07T20:46:10.064Z · LW · GW

Confound: I may also start eating a lot more collagen/gelatin, because it is delicious and afaict it does something.

Comment by mako yass (MakoYass) on Why you should be using a retinoid · 2024-10-06T03:43:23.768Z · LW · GW

My (34) skin has just now started to look aged. In response to that and migraines (linked to magnesium deficiency), I've started eating liver a lot. I'll report back in a year.

Comment by mako yass (MakoYass) on Eye contact is effortless when you’re no longer emotionally blocked on it · 2024-10-04T21:57:58.300Z · LW · GW

That addresses the concern.

Comment by mako yass (MakoYass) on Verifiable private execution of machine learning models with Risc0? · 2024-09-30T23:25:55.237Z · LW · GW

https://ezkl.xyz/ 👀

Comment by mako yass (MakoYass) on Eye contact is effortless when you’re no longer emotionally blocked on it · 2024-09-29T21:39:25.846Z · LW · GW

This can be quite a bad thing, since a person's face often tells you whether what you're saying is landing for them or whether you need to elaborate on certain points (unless they have a people pleaser complex, in which case they'll just nod and smile always even when they're confused and offended on the inside lmao). The worst I've seen it was this discussion with Avi Loeb where he was lecturing someone who he had a disagreement with and he actually closed his eyes while he was talking and although I'm sure it wasn't fully self-aware about it, it was very arrogant. He was not talking to that person; he must waste a lot of time, in reckonings, retreading old ground without making progress towards reconciliation.