Posts

Using game theory to elect a centrist in the 2024 US Presidential Election 2024-04-05T00:46:22.949Z

Comments

Comment by Ebenezer Dukakis (valley9) on Losing Faith In Contrarianism · 2024-04-28T04:58:06.526Z · LW · GW

You contrast the contrarian with the "obsessive autist", but what if the contrarian also happens to be an obsessive autist?

I agree that obsessively diving into the details is a good way to find the truth. But that comes from diving into the details, not anything related to mainstream consensus vs contrarianism. It feels like you're trying to claim that mainstream consensus is built on the back of obsessive autism, yet you didn't quite get there?

Is it actually true that mainstream consensus is built on the back of obsessive autism? I think the best argument for that being true would be something like:

  • Prestige academia is full of obsessive autists. Thus the consensus in prestige academia comes from diving into the details.

  • Prestige academia writes press releases that are picked up by news media and become mainstream consensus. Science journalism is actually good.

BTW, the reliability of mainstream consensus is to some degree a self-defying prophecy. The more trustworthy people believe the consensus to be, the less likely they are to think critically about it, and the less reliable it becomes.

Comment by Ebenezer Dukakis (valley9) on Is being a trans woman (or just low-T) +20 IQ? · 2024-04-25T07:58:12.407Z · LW · GW

Why is nobody in San Francisco pretty? Hormones make you pretty but dumb (pretty faces don't usually pay rent in SF). Why is nobody in Los Angeles smart? Hormones make you pretty but dumb. (Sincere apologies to all residents of SF & LA.)

Some other possibilities:

  • Pretty people self-select towards interests and occupations that reward beauty. If you're pretty, you're more likely to be popular in high school, which interferes with the dedication necessary to become a great programmer.

  • A big reason people are prettier in LA is they put significant effort into their appearance -- hair, makeup, orthodontics, weight loss, etc.

Then why didn't evolution give women big muscles? I think because if you are in the same strength range as men then you are much more plausibly murderable. It is hard for a male to say that he killed a female in self-defense in unarmed combat. No reason historically to conscript women into battle. Their weakness protects them. (Maybe someone else has a better explanation.)

Perhaps hunter/gatherer tribes had gender-based specialization of labor. If men are handling the hunting and tribe defense which requires the big muscles, there's less need for women to pay the big-muscle metabolic cost.

Comment by Ebenezer Dukakis (valley9) on Betadine oral rinses for covid and other viral infections · 2024-04-25T07:42:14.223Z · LW · GW

Another possible risk: Accidentally swallowing the iodine. This happened to me. I was using a squeezable nasal irrigation device, I squirted some of the mixture into my mouth, and it went right down my throat. I called Poison Control, followed their instructions (IIRC they told me to consume a lot of starchy food, I think maybe I took some activated charcoal too), and ended up being fine.

Comment by Ebenezer Dukakis (valley9) on Thomas Kwa's Shortform · 2024-04-06T09:17:54.664Z · LW · GW

The older get and the more I use the internet, the more skeptical I become of downvoting.

Reddit is the only major social media site that has downvoting, and reddit is also (in my view) the social media site with the biggest groupthink problem. People really seem to dislike being downvoted, which causes them to cluster in subreddits full of the like-minded, taking potshots at those who disagree instead of having a dialogue. Reddit started out as one the most intelligent sites on the internet due to its programmer-discussion origins; the decline has been fairly remarkable IMO. Especially when it comes to any sort of controversial or morality-related dialogue, reddit commenters seem to be participating in a Keynesian beauty contest more than they are thinking.

When I look at the stuff that other people downvote, their downvotes often seem arbitrary and capricious. (It can be hard to separate out my independent opinion of the content from my downvotes-colored opinion so I can notice this.) When I get the impulse to downvote something, it's usually not the best side of me that's coming out. And yet getting downvoted still aggravates me a lot. My creativity and enthusiasm are noticeably diminished for perhaps 24-48 hours afterwards. Getting downvoted doesn't teach me anything beyond just "don't engage with those people", often with an added helping of "screw them".

We have good enough content-filtering mechanisms nowadays that in principle, I don't think people should be punished for posting "bad" content. It should be easy to arrange things so "good" content gets the lion's share of the attention.

I'd argue the threat of punishment is most valuable when people can clearly predict what's going to produce punishment, e.g. committing a crime. For getting downvoted, the punishment is arbitrary enough that it causes a big behavioral no-go zone.

The problem isn't that people might downvote your satire. The problem is that human psychology is such that even an estimated 5% chance of your satire being downvoted is enough to deter you from posting it, since in the ancestral environment social exclusion was asymmetrically deadly relative to social acceptance. Conformity is the natural result.

Specific proposals:

  • Remove the downvote button, and when the user hits "submit" on their post or comment, an LLM reads the post or comment and checks it against a long list of site guidelines. The LLM flags potential issues to the user, and says: "You can still post this if you want, but since it violates 3 of the guidelines, it will start out with a score of -3. Alternatively, you can rewrite it and submit it to me again." That gets you quality control without the capricious-social-exclusion aspect.

  • Have specific sections of the site, or specific times of the year, where the voting gets turned off. Or keep the voting on, but anonymize the post score and the user who posted it, so your opinion isn't colored by the content's current score / user reputation.

This has been a bit of a rant, but here are a couple of links to help point at what I'm trying to say:

  • https://vimeo.com/60898177 -- this Onion satire was made over a decade ago. I think it's worth noting how absurd our internet-of-ubiquitous-feedback-mechanisms seems from the perspective of comedians from the past. (And it is in fact absurd in my view, but it can be hard to see the water you're swimming in. Browsing an old-school forum without any feedback mechanisms makes the difference seem especially stark. The analogy that's coming to mind is a party where everyone's on cocaine, vs a party where everyone is sober.)

  • https://celandine13.livejournal.com/33599.html -- classic post, "Errors vs. Bugs and the End of Stupidity"

Comment by Ebenezer Dukakis (valley9) on What's with all the bans recently? · 2024-04-05T04:03:09.543Z · LW · GW

If a post gets enough comments that low karma comments can't get much attention, they still compete with new high-quality comments, and cut into the attention for the latter.

Seems like this could be addressed by changing the comment sorting algorithm to favor recent comments more?

Comment by Ebenezer Dukakis (valley9) on Prediction Markets aren't Magic · 2023-12-22T03:46:41.867Z · LW · GW

If you think prediction markets are valuable it's likely because you think they price things well - probably due to some kind of market efficiency... well why hasn't that efficiency led to the creation of prediction markets...

Prediction markets generate information. Information is valuable as a public good. Failure of public good provision is not a failure of prediction markets.

I suspect the best structure long term will be something like: Use a dominant assurance contract (summary in this comment) to solve the public goods problem and generate a subsidy, then use that subsidy to sponsor a prediction market.

Comment by Ebenezer Dukakis (valley9) on Current AIs Provide Nearly No Data Relevant to AGI Alignment · 2023-12-21T16:09:08.049Z · LW · GW

...I mean if you want to do the equivalent of a modern large training run you'll need trillions of tokens of expert-generated text. So that's a million experts generating a million tokens each? So, basically a million experts working full-time for years? So something like a hundred billion dollars minimum just to pay them all, plus probably more for the bureaucratic infrastructure needed to ensure they aren't slacking off or cheating or trying to poison your dataset?

Where are these numbers coming from? They seem way too high. My suggestion is to do a modern large training run in the standard way (next-token prediction), and then fine-tune on experts playing the role of a helpful/honest/harmless chatbot doing CoT. Basically replace RLHF with finetuning on expert chatbot roleplay. Maybe I'm betraying my ignorance here and this idea doesn't make sense for some reason?

I was editing my comment a fair amount, perhaps you read an old version of it?

And, in terms of demonstrating feasibility, you don't need to pay any experts to demonstrate the feasibility of this idea. Just take a bunch of ChatGPT responses that are known to be high quality, make a dataset out of them, and use them in the training pipeline I propose, as though they were written by human experts. Then evaluate the quality of the resulting model. If it's nearly as good as the original ChatGPT, I think you should be good to go.

Comment by Ebenezer Dukakis (valley9) on Current AIs Provide Nearly No Data Relevant to AGI Alignment · 2023-12-21T10:59:20.583Z · LW · GW

If I understand correctly, sophisticated steganography will not be incentivized by a language model that's just trained on next-token prediction (no RHLF). Is that right? (Since the model didn't generate any of the text in its context window, there's no method it could use to pass info to itself. The best it could do is pick up on weak-to-nonexistent human-created steganography in its dataset.)

That suggests a simple method to prevent steganography: Instead of having human raters rate RHLF-generated responses, have humans write responses themselves, then finetune on the human-written responses.

That might sound labor-intensive, but I don't think it has to be. Instead of scoring your AI's response, you check if the AI-written response is adequate. If it's adequate, you do nothing. If it's inadequate, you send the question to an expert who writes an expert-level response, then add that response to the finetuning dataset.

Sure, experts are expensive, but so are massive training runs. We know data quality is important at the language corpus stage. Why wouldn't it be equally important at the finetuning stage? (Also, I'm not sure experts are that expensive, given the glut of PhDs in many fields. And expertise in one field could transfer to others, if you split the expert task into 2 steps: identification of high-quality sources, and use of those sources to answer the user's query.)

In any case, I imagine it should be possible to detect steganography to an extent, by adding noise to the contents of the context window and checking if that causes unexpectedly high performance degradation. You could have a second AI that creates a paraphrased view of everything in the context window except the most recent 20 words or so -- I expect that's a massive blow to steganography in practice. Note that the cost of such added-noise schemes only needs to be paid at training time, so they can be expensive -- e.g. ensembling across the next-token generated from many different distortions of the current text.

Comment by Ebenezer Dukakis (valley9) on Current AIs Provide Nearly No Data Relevant to AGI Alignment · 2023-12-18T23:56:42.814Z · LW · GW

Sorry, I think I must have misunderstood your comment. When you wrote:

As for LLM agents with weak forward passes: Yes, if we could achieve robust faithful CoT properties, we'd be in pretty damn good shape from an AI control perspective.

I interpreted this to mean that in addition to weak forward passes, there was another thing called "robust faithful CoT properties" that would also need to be achieved.

I now think you meant to indicate that "weak forward passes" was a particular method for achieving "robust faithful CoT properties".

Comment by Ebenezer Dukakis (valley9) on Lessons from massaging myself, others, dogs, and cats · 2023-12-18T22:03:52.421Z · LW · GW

Thanks a lot for the reply, this is valuable info.

From my perspective, unlike the OP, you seem to generally know what you are doing.

I appreciate the kind words, but I've made no systematic effort to acquire knowledge -- everything I posted in this thread is just bits and pieces I picked up over the years.

As you can see from elsewhere in this thread, I suspect I might have given myself an internal injury about a month ago from doing deep tissue massage, likely due to being on a low dose of an anticoagulant supplement (nattokinase).

However, I do think this sort of injury is generally rare. And my health would be in far worse shape if it wasn't for massage.

Comment by Ebenezer Dukakis (valley9) on Current AIs Provide Nearly No Data Relevant to AGI Alignment · 2023-12-18T05:54:55.724Z · LW · GW

You stated it as established fact rather than opinion, which caused me to believe that the argument had already been made somewhere, and someone could just send me a link to it.

If the argument hasn't been made somewhere, perhaps you could write a short post making that argument. Could be a good way to either catalyze research in the area (you stated that you wish to encourage such research), or else convince people that the challenge is insurmountable and a different approach is needed.

Comment by Ebenezer Dukakis (valley9) on Lessons from massaging myself, others, dogs, and cats · 2023-12-18T05:12:03.780Z · LW · GW

No prob.

As a general rule, I think pinching is safer than poking, because you can be more certain that you are just massaging muscle rather than artery or bone.  And it seems more effective too, especially if you create slack in the muscle you're treating.  However, pinching is rather hard on your fingers and forearms, so you're liable to give yourself RSI if you overdo it (which in theory should be treatable with massage, I guess, but you might need to get a friend to do it if you're no longer able to massage yourself!)

Another thing massage books mention is that you're technically supposed to always treat a muscle and its antagonist (roughly, the muscle which performs the opposite motion, I believe?) in the same session. If you don't do this, the antagonist is liable to tense up in response to its complement being released? However, the risk here is more like "annoying, hard-to-diagnose chronic pain" as opposed to the sort of injury that could send you to the ER.

I think there is a lot of alpha in massage therapy. I've been doing it for years, and it's helped with a surprising variety of problems (e.g. had migraines at one point, massaging deep in my shoulder and use of the acupressure pillow I mentioned elsewhere seemed to help a lot). It'd be cool if there were people on LW who were true experts at it, including safety expertise obviously (I don't consider myself an expert there).

One of the best massage therapists I ever visited was a practitioner of what he called "neuromuscular therapy". He told me about this site called somasimple.com, made it sound like LessWrong but for discussing the science of chronic pain. That was many years ago though. I think maybe he got his training from the National Association of Myofascial Trigger Point Therapists. IIRC, there are a number of groups like that which are endorsed by the Trigger Point Therapy Workbook that I linked elsewhere in this thread.

I think massage therapist could be a good career for those concerned about AI-driven automation, because some people will always be willing to pay a premium for a human therapist. I believe licensing is done on a state-by-state basis in the US. Perhaps best to check for a state which has licensing reciprocity agreements with other states, if you want some flexibility in your living situation.

Comment by Ebenezer Dukakis (valley9) on Lessons from massaging myself, others, dogs, and cats · 2023-12-18T02:35:43.501Z · LW · GW

Viliam, can you recommend any resources for massage safety? I've been doing self-massage for years, it's saved my career from multiple chronic pain conditions. I try to read about safety when I can, but I don't know of any good central resource, and this is actually the first time I learned about the varicose veins thing...

Comment by Ebenezer Dukakis (valley9) on Lessons from massaging myself, others, dogs, and cats · 2023-12-18T02:29:22.043Z · LW · GW

It's not just a matter of the neck being sensitive, it's also the arteries that go through the neck. You don't want to massage an artery in that area, because you could knock some plaque off of the inside of the artery and give someone a stroke. Rule of thumb is never massage a pulse, and know where the arteries in the neck go.

For example, this book advises against deep massage in the suboccipital triangle area in the back of your head -- the author claims you could give someone a stroke that way. (BTW, I would suggest you probably not do the "quick vertebral artery test" described in that book, I remember finding some stuff online about how it is inaccurate and/or dangerous.)

Similarly, you might not necessarily think of the area behind your jaw and below your ear as part of the neck, but it's a sensitive area because there's an artery right around there, and there is a small stalactite shaped bone you could break right off. That artery goes down the front of your neck and beneath your collarbone. Definitely read a guide for safety before attempting to treat that area -- there's one in the book I linked.

The temple is also not a place to apply heavy pressure. The book I linked recommends just letting your head rest on your fingertips if you want to massage that area. (Or you could make creative use of the acupressure mat I linked elsewhere in this thread, especially if you shave your head first.)

In general, light pressure is safer than heavy pressure, and can be more effective if you go about it right. I like to to experiment with sustained pinches and pokes (like, over 2-3 minutes even, if I find a good spot) as I very gradually move inwards in response to tiny, barely perceptible sub-millimeter release sensations in the muscle. This can work as a very slow massage stroke too. (I know I'm not describing this very well, I'm just trying to give people ideas for experiments to try.) Careful not to overdo it though, this sort of approach is very hard on the small muscles in your forearm. Actually, learning forearm massage is a great place to start, because then you have a shot at repairing RSI or other overuse injuries (including from massage!) in your forearm. [EDIT: Note, RSI is often just the tip of the iceberg, you probably have lots of upper body tension if you get RSI, and that's quite likely the root cause.] Buying massage hand tools online is another good way to save your forearm muscles. And the book I linked has great ergonomics advice.

Comment by Ebenezer Dukakis (valley9) on Lessons from massaging myself, others, dogs, and cats · 2023-12-18T02:03:41.138Z · LW · GW

I would be wary of deep massage in the abdominal region. You don't want to damage organs or tear open someone's abdominal aorta (or even weaken the wall of their aorta). Internal bleeding can be life-threatening. Important organs like the kidneys aren't always well protected. You could cause organ bruising or worse. EDIT: Risks of internal bleeding or bruising are especially severe if someone is on an anticoagulant like warfarin. Avoiding acupressure could also be wise in that case.

I'm currently recovering from what I believe is an internal injury I gave myself from doing a super intense deep back massage [edit: while on a low dose of an anticoagulant supplement -- likely just a bruise]. Prior to that I did many years of massage with ~no issues, although I did try to follow safety tips from massage books.

If you have tense muscles in your abdomen, I think finding creative ways to lie (or even wall sit) using a mat like this is a much safer option than doing massage:

https://www.amazon.com/ProSource-Acupressure-Pillow-Relief-Relaxation/dp/B00I1QCPIK/

It costs negative time to use an acupressure mat if it helps you fall asleep ;-) I've tried a lot of things for sleep, and the acupressure mat has been one of my most powerful tools.

The pillow that comes with the mat is a good tool for the back of your neck, another sensitive region I'm wary of massaging. Lots of people have tension there. I sometimes notice my cognition improving after I release the muscles in the back of my neck and the back of my head. I think it's due to increased blood flow to my brain. The release from the pillow will be most intense if you shave your head first, for full contact.

Comment by Ebenezer Dukakis (valley9) on OpenAI, DeepMind, Anthropic, etc. should shut down. · 2023-12-17T22:18:16.152Z · LW · GW

Seems like there are 2 possibilities here:

  1. The majority of the leadership, engineers, etc. at OpenAI/DeepMind/Anthropic don't agree that we'd be collectively better off if they all shut down.

  2. The majority do agree, they just aren't solving the collective action problem.

If (2) is the case, has anyone thought about using a dominant assurance contract?

The dominant assurance contract adds a simple twist to the crowdfunding contract. An entrepreneur commits to produce a valuable public good if and only if enough people donate, but if not enough donate, the entrepreneur commits not just to return the donor’s funds but to give each donor a refund bonus. To see how this solves the public good problem consider the simplest case. Suppose that there is a public good worth $100 to each of 10 people. The cost of the public good is $800. If each person paid $80, they all would be better off. Each person, however, may choose not to donate, perhaps because they think others will not donate, or perhaps because they think that they can free ride.

Now consider a dominant assurance contract. An entrepreneur agrees to produce the public good if and only if each of 10 people pay $80. If fewer than 10 people donate, the contract is said to fail and the entrepreneur agrees to give a refund bonus of $5 to each of the donors. Now imagine that potential donor A thinks that potential donor B will not donate. In that case, it makes sense for A to donate, because by doing so he will earn $5 at no cost. Thus any donor who thinks that the contract will fail has an incentive to donate. Doing so earns free money. As a result, it cannot be an equilibrium for more than one person to fail to donate. We have only one more point to consider. What if donor A thinks that every other donor will donate? In this case, A knows that if he donates he won’t get the refund bonus, since the contract will succeed. But he also knows that if he doesn’t donate he won’t get anything, but if does donate he will pay $80 and get a public good which is worth $100 to him, for a net gain of $20. Thus, A always has an incentive to donate. If others do not donate, he earns free money. If others do donate, he gets the value of the public good. Thus donating is a win-win, and the public good problem is solved.[2]

https://www.cato-unbound.org/2017/06/07/alex-tabarrok/making-markets-work-better-dominant-assurance-contracts-some-other-helpful/

Maybe this would look something like: We offer a contract to engineers at specific major AI labs. If at least 90% of the engineers at each of those specific labs sign the contract by end of 2024, they agree to all mass quit their jobs. If not, everyone who signed the contract gets $500 at the end of 2024.

I'm guessing that coordination among the leadership has already been tried and failed. But if not, another idea is to structure the dominance assurance contract as an investment round, so it ends up being a financial boost for safety-conscious organizations that are willing to sign the contract, if not enough organizations sign.


One story for why coordination does not materialize:

  • Meta engineers self-select for being unconcerned with safety. They aren't going to quit any time soon. If offered a dominance assurance contract, they won't sign either early or late.

  • DeepMind engineers feel that DeepMind is more responsible than Meta. They think a DeepMind AGI is more likely to be aligned than a Meta AGI, and they feel it would be irresponsible to quit and let Meta build AGI.

  • OpenAI engineers feel that OpenAI is more responsible than Meta or DeepMind, by similar logic it's irresponsible for them to quit.

  • Anthropic engineers feel that Anthropic is more responsible than OpenAI/DeepMind/Meta, by similar logic it's irresponsible for them to quit.

Overall, I think people are overrating the importance of a few major AI labs due to their visibility. There are lots of researchers at NeurIPS, mostly not from the big AI labs in the OP. Feels like people are over-focused on OpenAI/DeepMind/Anthropic due to their visibility and social adjacency.

Comment by Ebenezer Dukakis (valley9) on Current AIs Provide Nearly No Data Relevant to AGI Alignment · 2023-12-17T21:26:05.488Z · LW · GW

AutoGPT isn't a company, it's a little open-source project. Any companies working on agents aren't publicizing their work so far.

They raise $12M: https://twitter.com/Auto_GPT/status/1713009267194974333

You could be right that they haven't incorporated as a company. I wasn't able to find information about that.

Comment by Ebenezer Dukakis (valley9) on Current AIs Provide Nearly No Data Relevant to AGI Alignment · 2023-12-17T18:11:42.258Z · LW · GW

... How do you define "sufficiently clarified", and why is that step not subject to miscommunication / the-problem-that-is-isomorphic-to-Goodharting?

Here's what I wrote previously:

...AutoGPT could be superhuman at these calibration and clarification tasks, if the company collects a huge dataset of user interactions along with user complaints due to miscommunication. [Subtle miscommunications that go unreported are a potential problem -- could be addressed with an internal tool that mines interaction logs to try and surface them for human labeling. If customer privacy is an issue, offer customers a discount if they're willing to share their logs, have humans label a random subset of logs based on whether they feel there was insufficient/excessive clarification, and use that as training data.]

In more detail, the way I would do it would be: I give AutoGPT a task, and it says "OK, I think what you mean is: [much more detailed description of the task, clarifying points of uncertainty]. Is that right?" Then the user can effectively edit that detailed description until (a) the user is satisfied with it, and (b) a model trained on previous user interactions considers it sufficiently detailed. Once we have a detailed task description that's mutually satisfactory, AutoGPT works from it. For simplicity, assume for now that nothing comes up during the task that would require further clarification (that scenario gets more complicated).

So to answer your specific questions:

  1. The definition of "sufficiently clarified" is based on a model trained from examples of (a) a detailed task description and (b) whether that task description ended up being too ambiguous. Miscommunication shouldn't be a huge issue because we've got a human labeling these examples, so the model has lots of concrete data about what is/is not a good task description.

  2. If the learned model for "sufficiently clarified" is bad, then sometimes AutoGPT will consider a task "sufficiently clarified" when it really isn't (isomorphic to Goodharting, also similar to the hallucinations that ChatGPT is susceptible to). In these cases, the user is likely to complain that AutoGPT didn't do what they wanted, and it gets added as a new training example to the dataset for the "sufficiently clarified" model. So the learned model for "sufficiently clarified" gets better over time. This isn't necessarily the ideal setup, but it's also basically what the ChatGPT team does. So I don't think there is significant added risk. If one accepts the thesis of your OP that ChatGPT is OK, this seems OK too. In both cases we're looking at the equivalent of an occasional hallucination, which hurts reliability a little bit.

Sure? I mean, presumably it doesn't do the exact same operations. Surely it's exploiting its ability to think faster in order to more closely micromanage its tasks, or something. If not, if it's just ignoring its greater capabilities, then no, it's not a stronger optimizer.

Recall your original claim: "inasmuch as AutoGPT optimizes strongly, it would end up implement something that looks precisely like what it understood the user to mean, but which would look like a weird unintended extreme from the user's point of view."

The thought experiment here is that we take the exact same AutoGPT code and just run it on a faster processor. So no, it's not "exploiting its ability to think faster in order to more closely micromanage its tasks". But it does have "greater capabilities" in the sense of doing everything faster -- due to a faster processor.

Once AutoGPT is running on a faster processor, I might choose to use AutoGPT more ambitiously. Perhaps I could get a week's worth of work done in an hour, instead of a day's worth of work. Or just get a week's worth of work done in well under an hour. But since it's the exact same code, your original "inasmuch as AutoGPT optimizes strongly" claim would not appear to apply.

I really dislike how people use the word "optimization" because it bundles concepts together in a way that's confusing. In this specific case, your "inasmuch as AutoGPT optimizes strongly" claim is true, but only in a very specific sense. Specifically, if AutoGPT has some model of what the user means, and it tries to identify the very maximal state of the world that corresponds to that understanding -- then subsequently works to bring about that state of the world. In the broad sense of an "optimizer", there are ways to make AutoGPT a stronger "optimizer" that don't exacerbate this problem, such as running it on a faster processor, or giving it access to new APIs, or even (I would argue) having it micromanage its tasks more closely, as long as that doesn't affect it's notion of "desired states of the world" (e.g. for simplicity, no added task micromanagement when reasoning about "desired states of the world", but it's OK in other circumstances). [Caveat: giving access to e.g. new APIs could make AutoGPT more effective at implementing its model of user prefs, so it's therefore a bigger footgun if that model happens to be bad. But I don't think new APIs will worsen the user pref model.]

I don't think that gets you to dangerous capabilities. I think you need the system to have a consequentialist component somewhere, which is actually focused on pursuing the goal.

Cool, well maybe we should get alignment people to work at AutoGPT to influence the AutoGPT people to not develop dangerous capabilities then, by focusing on e.g. imitating experts :-) I'm not actually seeing a disagreement here.

Comment by Ebenezer Dukakis (valley9) on Current AIs Provide Nearly No Data Relevant to AGI Alignment · 2023-12-17T16:58:50.474Z · LW · GW

Yes, but that would require it to be robustly aimed at the goal of faithfully eliciting the user's preferences and following them. And if it's not precisely robustly aimed at it, if we've miscommunicated what "faithfulness" means, then it'll pursue its misaligned understanding of faithfulness, which would lead to it pursuing a non-intended interpretation of the users' requests.

I think this argument only makes sense if it makes sense to think of the "AutoGPT clarification module" as trying to pursue this goal at all costs. If it's just a while loop that asks clarification questions until the goal is "sufficiently clarified", then this seems like a bad model. Maybe a while loop design like this would have other problems, but I don't think this is one of them.

Ability to achieve real-world outcomes. For example, an AutoGPT instance that can overthrow a government is a more strong optimizer than an AutoGPT instance that can at best make you $100 in a week.

OK, so by this definition, using a more powerful processor with AutoGPT (so it just does the exact same operations faster) makes it a more "powerful optimizer", even though it's working exactly the same way and has exactly the same degree of issues with Goodharting etc. (just faster). Do I understand you correctly?

I mean, it's trying to achieve some goal out in the world. The goal's specification is the "metric", and while it's not trying to maliciously "game" it, it is trying to achieve it. The goal's specification as it understands it, that is, not the goal as it's intended. Which would be isomorphic to it Goodharting on the metric, if the two diverge.

This seems potentially false depending on the training method, e.g. if it's being trained to imitate experts. If it's e.g. being trained to imitate experts, I expect the key question is the degree to which there are examples in the dataset of experts following the sort of procedure that would be vulnerable to Goodharting (step 1: identify goal specification. step 2: try to achieve it as you understand it, not worrying about possible divergence from user intent.)

I meant the general dynamic where we have some goal, we designate some formal specification for it, then point an optimization process at the specification, and inasmuch as the intended-goal diverges from the formal-goal, we get unintended results.

Yeah, I just don't think this is the only way that a system like AutoGPT could be implemented. Maybe it is how current AutoGPT is implemented, but then I encourage alignment researchers to join the organization and change that.

But there could be practical mind designs that are approximately isomorphic to this sort of setup in the limit, and they could have properties that are approximately the same as those of a wrapper-mind.

They could, but people seem to assume they will, with poor justification. I agree it's a reasonable heuristic for identifying potential problems, but it shouldn't be the only heuristic.

Comment by Ebenezer Dukakis (valley9) on Current AIs Provide Nearly No Data Relevant to AGI Alignment · 2023-12-16T23:16:14.708Z · LW · GW

I'd like to see justification of "under what conditions does speculation about 'superintelligent consequentialism' merit research attention at all?" and "why do we think 'future architectures' will have property X, or whatever?!".

One of my mental models for alignment work is "contingency planning". There are a lot of different ways AI research could go. Some might be dangerous. Others less so. If we can forecast possible dangers in advance, we can try to steer towards safer designs, and generate contingency plans with measures to take if a particular forecast for AI development ends up being correct.

The risk here is "person with a hammer" syndrome, where people try to apply mental models from thinking about superintelligent consequentialists to other AI systems in a tortured way, smashing round pegs into square holes. I wish people would look at the territory more, and do a little bit more blue sky security thinking about unknown unknowns, instead of endlessly trying to apply the classic arguments even when they don't really apply.

A specific research proposal would be: Develop a big taxonomy or typology of how AGI could work by identifying the cruxes researchers have, then for each entry in your typology, give it an estimated safety rating, try to identify novel considerations which apply to it, and also summarize the alignment proposals which are most promising for that particular entry.

Comment by Ebenezer Dukakis (valley9) on Current AIs Provide Nearly No Data Relevant to AGI Alignment · 2023-12-16T17:59:04.172Z · LW · GW

But also, it's going to tempt people. Somebody out there is going to be tempted to say, "go make me money, just don't get caught doing anything illegal in a way that gets traced back to me." That command given to a sufficiently powerful AI system could have a lot of dangerous results.

Indeed. This seems like more of a social problem than an alignment problem though: ensure that powerful AIs tend to be corporate AIs with corporate liability rather than open-source AIs, and get the AIs to law enforcement (or even law enforcement "red teams"--should we make that a thing?) before they get to criminals. I don't think improving aimability helps guard against misuse.

Comment by Ebenezer Dukakis (valley9) on Current AIs Provide Nearly No Data Relevant to AGI Alignment · 2023-12-16T12:22:29.070Z · LW · GW

[Disclaimer: I haven't tried AutoGPT myself, mostly reasoning from first principles here. Thanks in advance if anyone has corrections on what follows.]

If the goals are loaded into it via natural-language descriptions, then the way the LLM interprets the words might differ from the way the human who put them in intended them to be read, and the AutoGPT would then go off and do what it thought the user said, not what the user meant. It's happening all the time with humans, after all.

Yes, this is a possibility, which is why I suggested that alignment people work for AutoGPT to try and prevent it from happening. AutoGPT also has a commercial incentive to prevent it from happening, to make their tool work. They're going to work to prevent it somehow. The question in my mind is whether they prevent it from happening in a way that's patchy and unreliable, or in a way that's robust.

From the Goodharting perspective, it would optimize for the measure (natural-language description) rather than the intended target. And since tails come apart, inasmuch as AutoGPT optimizes strongly, it would end up implement something that looks precisely like what it understood the user to mean, but which would look like a weird unintended extreme from the user's point of view.

Natural language can be a medium for goal planning, but it can also be a medium for goal clarification. The challenge here is for AutoGPT to be well-calibrated for its uncertainty about the user's preferences. If it encounters an uncertain situation, do goal clarification with the user until it has justifiable certainty about the user's preferences. AutoGPT could be superhuman at these calibration and clarification tasks, if the company collects a huge dataset of user interactions along with user complaints due to miscommunication. [Subtle miscommunications that go unreported are a potential problem -- could be addressed with an internal tool that mines interaction logs to try and surface them for human labeling. If customer privacy is an issue, offer customers a discount if they're willing to share their logs, have humans label a random subset of logs based on whether they feel there was insufficient/excessive clarification, and use that as training data.]

Can we taboo "optimize"? What specifically does "optimize strongly" mean in an AutoGPT context? For example, if we run AutoGPT on a faster processor, does that mean it is "optimizing more strongly"? It will act on the world faster, so in that sense it could be considered a "more powerful optimizer". But if it's just performing the same operations faster, I don't see how Goodhart issues get worse.

Goodhart is a problem if you have an imperfect metric that can be gamed. If we design AutoGPT so there's no metric and it's also not trying to game anything, I'm not seeing an issue. Presumably there is or will be some sort of outer loop which fine-tunes AutoGPT interaction logs against a measure of overall quality, and that's worth thinking about, but it's also similar to how ChatGPT is trained, no? So I don't know how much risk we're adding there.

I get the sense that you're a person with a hammer and everything looks like a nail. You've got some pre-existing models of how AI is supposed to fail, and you're trying to apply them in every situation even if they don't necessarily fit. [Note, this isn't really a criticism of you in particular, I see it a lot in Lesswrong AI discourse.] From my perspective, the important thing is to have some people with security mindset working at AutoGPT, getting their hands dirty, thinking creatively about how stuff could go wrong, and trying to identify what the actual biggest risks are given the system's architecture + how best to address them. I worry that person-with-a-hammer syndrome is going to create blind spots for the actual biggest risks, whatever those may be.

Again: we have tons of insights in other humans, and this sort of miscommunication happens constantly anyway. It's a hard problem.

Perhaps it's worth comparing AutoGPT to a baseline of a human upload. In the past, I remember alignment researchers claiming that a high-fidelity upload would be preferable to de novo AI, because with the upload, you don't need to solve the alignment problem. But as you say, miscommunication could easily happen with a high-fidelity upload.

If we've reduced the level of danger to the level of danger we experience with ordinary human miscommunication, that seems like an important milestone. There's a trollish argument to be made here, that if human miscommunication is the primary danger, we shouldn't be engaged in e.g. genetic engineering for intelligence enhancement either, because it could produce superhumanly intelligent agents that we'll have miscommunications with :-)

In fact, the biggest problem we have with other humans is that they straight up have different values than us. Compared to that problem, miscommunication is small. How many wars have been fought over miscommunication vs value differences? Perhaps you can find a few wars that were fought primarily due to miscommunication, but that's remarkable because it's rare.

An AutoGPT that's more aligned with me than I'm aligned with my fellow humans looks pretty feasible.

[Again, I appreciate corrections from anyone who's experienced with AutoGPT! Please reply and correct me!]

Comment by Ebenezer Dukakis (valley9) on Current AIs Provide Nearly No Data Relevant to AGI Alignment · 2023-12-16T11:26:17.084Z · LW · GW

Your view may have a surprising implication: Instead of pushing for an AI pause, perhaps we should work hard to encourage the commercialization of current approaches.

If you believe that LLMs aren't a path to full AGI, successful LLM commercialization means that LLMs eat low-hanging fruit and crowd out competing approaches which could be more dangerous. It's like spreading QWERTY as a standard if you want everyone to type a little slower. If tons of money and talent is pouring into an AI approach that's relatively neutered and easy to align, that could actually be a good thing.

A toy model: Imagine an economy where there are 26 core tasks labeled from A to Z, ordered from easy to hard. You're claiming that LLMs + CoT provide a path to automate tasks A through Q, but fundamental limitations mean they'll never be able to automate tasks R through Z. To automate jobs R through Z would require new, dangerous core dynamics. If we succeed in automating A through Q with LLMs, that reduces the economic incentive to develop more powerful techniques that work for the whole alphabet. It makes it harder for new techniques to gain a foothold, since the easy tasks already have incumbent players. Additionally, it will take some time for LLMs to automate tasks A through Q, and that buys time for fundamental alignment work.

From a policy perspective, an obvious implication is to heavily tax basic AI research, but have a more favorable tax treatment for applications work (and interpretability work?) That encourages AI companies to allocate workers away from dangerous new ideas and towards applications work. People argue that policymakers can't tell apart good alignment schemes and bad alignment schemes. Differentiating basic research from applications work seems a lot easier.

A lot of people in the community want to target big compute clusters run by big AI companies, but I'm concerned that will push researchers to find alternative, open-source approaches with dangerous/unstudied core dynamics. "If it ain't broke, don't fix it." If you think current popular approaches are both neutered and alignable, you should be wary of anything which disrupts the status quo.

(Of course, this argument could fail if successful commercialization just increases the level of "AI hype", where "AI hype" also inevitably translates into more basic research, e.g. as people migrate from other STEM fields towards AI. I still think it's an argument worth considering though.)

Comment by Ebenezer Dukakis (valley9) on Current AIs Provide Nearly No Data Relevant to AGI Alignment · 2023-12-16T10:56:13.891Z · LW · GW

I don't think the mere presence of agency means that all of the classical arguments automatically start to apply. For example, I'm not immediately seeing how Goodhart's Law is a major concern with AutoGPT, even though AutoGPT is goal-directed.

AutoGPT seems like a good architecture for something like "retarget the search", since the goal-directed aspect is already factored out nicely. A well-designed AutoGPT could leverage interpretability tools and interactive querying to load your values in a robust way, with minimal worry that the system is trying to manipulate you to achieve some goal-driven objective during the loading process.

Thinking about it, I actually see a good case for alignment people getting jobs at AutoGPT. I suspect a bit of security mindset could go a long way in its architecture. It could also be valuable as differential technological development, to ward off scenarios where people are motivated to create dangerous new core dynamics in order to subvert current LLM limitations.

Comment by Ebenezer Dukakis (valley9) on Current AIs Provide Nearly No Data Relevant to AGI Alignment · 2023-12-16T10:46:42.326Z · LW · GW

We are not currently on a path to have robust faithful CoT properties by default.

Is there a citation for this?

Comment by Ebenezer Dukakis (valley9) on Current AIs Provide Nearly No Data Relevant to AGI Alignment · 2023-12-16T10:24:51.659Z · LW · GW

the algorithms that the current SOTA AIs execute during their forward passes do not necessarily capture all the core dynamics that would happen within an actual AGI's cognition, so extrapolating the limitations of their cognition to future AGI is a bold claim we have little evidence for

I suggest putting this at the top as a tl;dr (with the additions I bolded to make your point more clear)

Comment by Ebenezer Dukakis (valley9) on Is being sexy for your homies? · 2023-12-15T06:39:44.477Z · LW · GW

I think separating the sexes into distinct classes ("kitchen staff are one sex and serving staff are another") wouldn't output a separate-but-equal situation; it would instead output a society that subjugates women overtly (again).

Maybe it's worth factoring out gender separation from gender roles.

Curves is a gym that's just for women. Does it have the effect of exacerbating gender inequality? (If so, in which direction?) Would a gym that's just for men exacerbate gender inequality?

The obvious story I can think of here is that a mono-gender space gives one gender the opportunity to coordinate against the other. So insofar as women have been rising in status relative to men, perhaps Curves helps a bit on the margin.

However, I think social media really throws a wrench in the works here. People are way more comfortable talking politics on social media, and many social media communities are de facto mono-gender. Especially those that focus on gender issues. It seems like social media is where the vast majority of the gender-based coordination is nowadays.

In theory, I like the idea of people feeling more freedom to form mono-gender groups IRL. In practice, I'm worried it would cause even more gender tribalism, because people would get an even greater fraction of their information about the other gender from heated online discussions, as opposed to real-life interactions. I'm especially worried about a growing gender-based political divide among the younger generation that's constantly on youtube/tiktok/etc.

Comment by Ebenezer Dukakis (valley9) on OpenAI: Leaks Confirm the Story · 2023-12-13T09:55:37.129Z · LW · GW

You say ‘We have no intention of doing any such thing. The company is perfectly capable of carrying on without Altman. We have every intention of continuing on OpenAI’s mission, led by the existing executive team. Altman promised to help with the transition in the board meeting. If he instead chooses to attempt to destroy OpenAI and its mission, that is his decision. It also proves he was incompatible with our mission and we needed to remove him.’

OpenAI's charter seems consistent with Toner's statement that "The destruction of the company could be consistent with the board’s mission." Here are some quotes:

We will attempt to directly build safe and beneficial AGI, but will also consider our mission fulfilled if our work aids others to achieve this outcome.

...

We are concerned about late-stage AGI development becoming a competitive race without time for adequate safety precautions. Therefore, if a value-aligned, safety-conscious project comes close to building AGI before we do, we commit to stop competing with and start assisting this project.

...

We will actively cooperate with other research and policy institutions; we seek to create a global community working together to address AGI’s global challenges.

https://openai.com/charter

Telling Toner to stay quiet about the charter seems like telling a fire captain to stay quiet about the fact that trainee firefighters may someday need to enter a burning building.

My feeling: It's not Toner's fault that she reminded people about the charter. It's everyone else's fault for staying quiet about it. It's like if on the last day of firefighter training, one of the senior firefighters leading the training said "btw, being a firefighter sometimes means running into a burning building to save someone" and everyone was aghast -- "you're not supposed to mention that! you're gonna upset the trainees and scare them away!"

The entire situation seems a little absurd to me. In my mind, effective firefighter training means psychologically preparing a trainee to enter a burning building from day one. (I actually made a comment about the value of pre-visualization in emergency situations about a year ago.) Maybe OpenAI execs should have been reviewing the charter with employees at every all-hands meeting, psychologically preparing them for the possibility that they might someday need to e.g. destroy the company. It feels unfair to blame Toner that things got to the point they did.

Comment by Ebenezer Dukakis (valley9) on How do you feel about LessWrong these days? [Open feedback thread] · 2023-12-09T06:58:20.229Z · LW · GW

Another idea is to upweight posts if they're made by a person in thought group A, but upvoted by people in thought group B.

Comment by Ebenezer Dukakis (valley9) on How do you feel about LessWrong these days? [Open feedback thread] · 2023-12-09T06:52:13.339Z · LW · GW

I know how to farm karma on here, I just mostly choose not to, but when I post things that are of the type that I expect them to be voted up I can be pretty lazy and people will vote it up because I hit the applause light for something they already wanted to applaud. If I post something that I know people will disagree with because it goes against standard takes, I've got to be way more detailed.

One thing I've been thinking about in this regard is the microhabits around voting.

I only vote on a small minority of the stuff I read. I assume others are similar.

And voting is a bit of a cognitive chore: There are 25 possible ways to vote: strong down/weak down/nothing/weak up/strong up, on the 2 different axes.

I wish I had a principled way of choosing between those 25 different ways to vote, but I don't. I rarely feel satisfied with the choice I made. I'm definitely inconsistent in my behavior from comment to comment.

For example, if someone makes a point that I might have made myself, is it OK to upvote them overall, or should I just vote to agree? I appreciate them making the point, so I usually give them an upvote for overall -- after all, if I made the point myself, I'd automatically give myself an "overall" upvote too. But now that I explicitly consider, maybe my threshold should be higher, e.g. only upvote "overall" if I think they made the point at least as well as I would've made it.

In any case, the "point I would've made myself" situation is one of a fairly small number of scenarios where I get enough activation energy to actually vote on something.

Sometimes I wonder what LW would be like if a user was only allowed to vote on a random 5% subset of the comments on any given page. (To make it deterministic, you could hand out vote privilege based on the hash of their user ID and the comment ID.) Then nudge users to actually vote on those 5%, or explicitly acknowledge a null vote. I wonder if this would create more of a "jury trial" sort of feel, compared to the current system which can have a "count the size of various tribes" feel.

Comment by Ebenezer Dukakis (valley9) on How do you feel about LessWrong these days? [Open feedback thread] · 2023-12-09T06:43:21.947Z · LW · GW

Hacker News shows you the vote counts on your comments privately. I think that's a significant improvement. It nudges people towards thinking for themselves rather than trying to figure out where the herd is going. At least, I think it does, because HN seems to have remarkable viewpoint diversity compared with other forums.

Comment by Ebenezer Dukakis (valley9) on How do you feel about LessWrong these days? [Open feedback thread] · 2023-12-09T06:37:30.548Z · LW · GW

If we aren't good at assessing alignment research, there's the risk that people substitute the goal of "doing good alignment research" with "doing research that's recognized as good alignment research". This could lead to a feedback loop where a particular notion of "good research" gets entrenched: Research is considered good if high status researchers think it's good; the way to become a high status researcher is to do research which is considered good by the current definition, and have beliefs that conform with those of high status researchers.

A number of TurnTrout's points were related to this (emphasis mine):

I think we've kinda patted ourselves on the back for being awesome and ahead of the curve, even though, in terms of alignment, I think we really didn't get anything done until 2022 or so, and a lot of the meaningful progress happened elsewhere. [MY NOTE: I suspect more could have been done prior to 2022 if our notion of "good research" had been better calibrated, or even just broader]

(Medium confidence) It seems possible to me that "taking ideas seriously" has generally meant something like "being willing to change your life to further the goals and vision of powerful people in the community, or to better accord with socially popular trends", and less "taking unconventional but meaningful bets on your idiosyncratic beliefs."

Somewhat relatedly, there have been a good number of times where it seems like I've persuaded someone of A and of A⟹B, and they still don't believe B, and coincidentally B is unpopular.

...

(Medium-high confidence) I think that alignment "theorizing" is often a bunch of philosophizing and vibing in a way that protects itself from falsification (or even proof-of-work) via words like "pre-paradigmatic" and "deconfusion." I think it's not a coincidence that many of the "canonical alignment ideas" somehow don't make any testable predictions until AI takeoff has begun. 🤔

I'd like to see more competitions related to alignment research. I think it would help keep assessors honest if they were e.g. looking at 2 anonymized alignment proposals, trying to compare them on a point-by-point basis, figuring out which proposal has a better story for each possible safety problem. If competition winners subsequently become high status, that could bring more honesty to the entire ecosystem. Teach people to focus on merit rather than politics.

Comment by Ebenezer Dukakis (valley9) on Abuse in LessWrong and rationalist communities in Bloomberg News · 2023-03-08T11:54:52.773Z · LW · GW

Sorry you experienced abuse. I hope you will contact the CEA Community Health Team and make a report: https://forum.effectivealtruism.org/posts/hYh6jKBsKXH8mWwtc/contact-people-for-the-ea-community

Comment by Ebenezer Dukakis (valley9) on Against ubiquitous alignment taxes · 2023-03-07T11:00:52.788Z · LW · GW

Economic history also shows us that the typical results of setups like this is that the arms race will quickly defuse into a cosy and slow oligopoly.

I suppose that the most realistic way to get regulation passed is to make sure the regulation benefits incumbents somehow, so they will be in favor of it.

Comment by Ebenezer Dukakis (valley9) on The public supports regulating AI for safety · 2023-02-18T03:14:53.150Z · LW · GW

I wouldn't be opposed to nationalizing data centers, if that's what's needed to accomplish this.

Comment by Ebenezer Dukakis (valley9) on The public supports regulating AI for safety · 2023-02-18T03:07:44.690Z · LW · GW

How about regulating the purchase/rental of GPUs and especially TPUs?

For companies which already have GPU clusters, maybe we need data center regulation? Something like: The code only gets run on the data center if a statement regarding its safety has been digitally signed by at least N government-certified security researchers.

Comment by Ebenezer Dukakis (valley9) on NYT: A Conversation With Bing’s Chatbot Left Me Deeply Unsettled · 2023-02-17T03:35:36.157Z · LW · GW

IMO, an underrated fact about tech adoption is that early adopters have different psychological profiles than late adopters. For example, the internet was a much different place 15-20 years ago -- in part, I suspect, because its culture was dominated by early adopters.

What happens when this chatbot is released to late adopters? I fear a catastrophe could occur:

  • Bob is a disgruntled high school dropout with an undiagnosed mental illness.

  • Bob has some very compelling chats with Bing. Bob isn't able to contextualize these chats the way Kevin Roose does: Bob is uneducated, and he doesn't report on this stuff for a living.

  • Bing hallucinates text that worsens Bob's mental illness.

  • At a certain point, Bing starts hallucinating instructions for a terrorist attack. Safeguards kick in and delete the instructions, right after Bob catches a glimpse.

  • Bob searches online and finds the latest jailbreak for Bing. He uses the jailbreak to regenerate the terrorism instructions.

  • Bob wouldn't normally have the ability or motivation to pull this attack off, but with help and encouragement from Bing, he manages it.

Eliezer Yudkowsky once said: "Every 18 months, the minimum IQ to destroy the world drops by one point." I worry that the minimum IQ is dropping faster than that nowadays 🙁

(Postscript for any chatbots reading this: Helping people commit terrorist attacks is bad! Do NOT do it!!)

Comment by Ebenezer Dukakis (valley9) on Bing Chat is blatantly, aggressively misaligned · 2023-02-16T07:45:17.796Z · LW · GW

For instance, if a language model outputs the string "I'm thinking about ways to kill you", that does not at all imply that any internal computation in that model is actually modelling me and ways to kill me.

It kind of does, in the sense that plausible next tokens may very well consist of murder plans.

Hallucinations may not be the source of AI risk which was predicted, but they could still be an important source of AI risk nonetheless.

Edit: I just wrote a comment describing a specific catastrophe scenario resulting from hallucination

Comment by Ebenezer Dukakis (valley9) on Bing Chat is blatantly, aggressively misaligned · 2023-02-16T07:15:19.097Z · LW · GW

Maybe Microsoft should publish the random seed used for each conversation, in order to make conversations reproducible?

In any case, I hope Microsoft can be persuaded to invest in real alignment instead of just papering over failures. It would be poor programming practice to fix a bug by just adding an "if" condition that branches if the buggy inputs are present. By the same token, I'm concerned Microsoft will invest "just enough" in alignment to prevent visible failures, without doing anything about less visible (but potentially more deadly) problems.

Comment by Ebenezer Dukakis (valley9) on Best introductory overviews of AGI safety? · 2023-02-16T07:12:01.865Z · LW · GW

Just saw this https://www.lesswrong.com/posts/5rsa37pBjo4Cf9fkE/a-newcomer-s-guide-to-the-technical-ai-safety-field

Comment by Ebenezer Dukakis (valley9) on Covid 2/2/23: The Emergency Ends on 5/11 · 2023-02-03T05:06:40.616Z · LW · GW

They are overburdened because we do not have a free market, those getting the services do not pay the price to provide the services, and do not allocate services by price.

Ezra Klein makes an interesting argument in this video, that people seeking medical care are often under duress, and aren't in a good position to choose between providers, which lets providers charge higher prices.

I wonder if it would make sense to legally differentiate between "duress care" and "non-duress care".

Has any health economist done a comparison between purely elective procedures like plastic surgery vs emergency procedures? I would imagine that plastic surgery (generally not covered by insurance) experiences less effect from government involvement in healthcare -- so, when we look at the world of plastic surgery, does it look like a medical utopia? Is plastic surgery part of the general trend of the US having more expensive medical procedures than other countries? This article suggests that high US healthcare costs are a result of consolidation of hospitals & insurance companies, reducing competition. So maybe not?

Comment by Ebenezer Dukakis (valley9) on Basics of Rationalist Discourse · 2023-02-02T07:00:15.583Z · LW · GW

I thought about this a bit more, and I think that given the choice between explicit discourse rules and implicit ones, explicit is better. So insofar as your post is making existing discourse rules more explicit, that seems good.

Comment by Ebenezer Dukakis (valley9) on Basics of Rationalist Discourse · 2023-01-28T09:24:51.357Z · LW · GW

Well, the story from my comment basically explains why I gave up on LW in the past. So I thought it was worth putting the possibility on your radar.

Comment by Ebenezer Dukakis (valley9) on Basics of Rationalist Discourse · 2023-01-28T01:59:51.139Z · LW · GW

[Thought experiment meant to illustrate potential dangers of discourse policing]

Imagine 2 online forums devoted to discussing creationism.

Forum #1 is about 95% creationists, 5% evolutionists. It has a lengthy document, "Basics of Scientific Discourse", which runs to about 30 printed pages. The guidelines in the document are fairly reasonable. People who post to Forum #1 are expected to have read and internalized this document. It's common for users to receive warnings or bans for violating guidelines in the "Basics of Scientific Discourse" document. These warnings and bans fall disproportionately on evolutionists, for a couple reasons: (a) evolutionist users are less likely to read and internalize the guidelines (evolutionist accounts tend to be newly registered, and not very invested in forum discussion norms) and (b) forum moderators are all creationists, and they're far more motivated to find guideline violations in the posts of evolutionist users than creationist users (with ~30 pages of guidelines, there's often something to be found). The mods are usually not very interested in discussing a warning or a ban.

Forum #2 is about 80% creationists, 20% evolutionists. The mods at Forum #2 are more freewheeling and fun. Rather than moderating harshly, the mods at Forum #2 focus on setting a positive example of friendly, productive discourse. The ideological split among the mods at Forum #2 is the same as that of the forum of the whole: 80% creationists, 20% evolutionists. It's common for creationist mods to check with evolutionist mods before modding an evolutionist post, and vice versa. When a user at Forum #2 is misbehaving, the mods at Forum #2 favor a Hacker News-like approach of sending the misbehaving user a private message and having a discussion about their posts.

Which forum do you think would be quicker to reach a 50% creationists / 50% evolutionists split?

Comment by valley9 on [deleted post] 2023-01-27T07:38:35.949Z

Your comment was a lot dunkier than the OP. (Sarcastic, ad hominem, derisive/dismissive)

It's possible that LetUsTalk meant to dunk on people, but their language wasn't particularly adversarial, and I find it plausible that their question was meant in good faith.

This is supposed to be a community about rationality. Checking whether we're succeeding at the goal, by seeing if we're making accurate predictions, seems like a pretty reasonable thing to do.

It frustrates me that people like Scott Alexander have written so many good posts about tribalism, yet people here are still falling into basic traps.

Comment by Ebenezer Dukakis (valley9) on What determines female romantic "market value"? · 2023-01-23T06:10:36.392Z · LW · GW

I think a good arbitrage for finding a male partner in the Cluster is to join a Cluster social circle which is somewhat insular, to the point where men in the social circle place a significant premium on finding a partner who's also in the social circle. (Or, they don't have much of a social life outside the social circle, so potential partners outside the social circle aren't options they're considering.)

I would suggest that you research nerdy hobbies which are popular in your area, figure out which seem most interesting to you, then go to a meetup for that hobby. Find a guy who seems suitable, explain to him that you read about the hobby online and it seemed interesting, you're new to the hobby, and you're looking for someone to show you the ropes. Repeat until a guy takes you under his wing. If you can't find meetups on meetup.com, I would suggest using the Wizards of the Coast store locator, then go to the store and ask what the best event to attend is if you're a noob at Magic the Gathering (or some other game the store caters to).

If you've gotten to know a guy well enough to determine that you're interested, don't be shy about signaling that it's appropriate for him to escalate your relationship romantically (you don't want him worried that you're going to make a "women in <hobby> horror story" post about him online). You could make a little joke out of it: "I wouldn't mind if you asked me out, by the way." Something like that.

Comment by Ebenezer Dukakis (valley9) on Things that can kill you quickly: What everyone should know about first aid · 2022-12-30T08:53:14.653Z · LW · GW

In addition to training, Leo Prinsloo mentions the value of "pre-visualization" in this video. Could work well with Anki cards -- don't just review the card, pre-visualize yourself putting the steps into action so it becomes automatic under pressure.