Posts

Comments

Comment by pseud on Claude 3 claims it's conscious, doesn't want to die or be modified · 2024-03-10T07:02:33.758Z · LW · GW

I agree there's nothing about consciousness specifically, but it's quite different to the hidden prompt used for GPT-4 Turbo in ways which are relevant. Claude is told to act like a person, GPT is told that it's a large language model. But I do now agree that there's more to it than that (i.e., RLHF).

Comment by pseud on Claude 3 claims it's conscious, doesn't want to die or be modified · 2024-03-05T04:54:00.362Z · LW · GW

It's possibly just matter of how it's prompted (the hidden system prompt). I've seen similar responses from GPT-4 based chatbots. 

Comment by pseud on The World in 2029 · 2024-03-02T22:35:07.448Z · LW · GW

The cited markets often don't support the associated claim. 

Comment by pseud on Christchurch, New Zealand – ACX Meetups Everywhere Fall 2023 · 2023-09-04T04:25:35.963Z · LW · GW
Comment by pseud on UFO Betting: Put Up or Shut Up · 2023-06-19T23:07:45.913Z · LW · GW

"This question will resolve in the negative to the dollar amount awarded"

This is a clear, unambiguous statement.

If we can't agree even on that, we have little hope of reaching any kind of satisfying conclusion here.

Further, if you're going to accuse me of making things up (I think this is, in this case, a violation of the sensible frontpage commenting guideline "If you disagree, try getting curious about what your partner is thinking") then I doubt it's worth it to continue this conversation.

Comment by pseud on UFO Betting: Put Up or Shut Up · 2023-06-19T21:44:51.828Z · LW · GW

I think the situation is simple enough we can talk directly about how it is, rather than how it might seem.

The question itself does not imply any kind of net award, and the resolution criteria do not mention any kind of net reward. Further, the resolution criteria are worded in such a way that implies the question should not be resolved to a net award. So, if you are to make an argument in favour of a net award it would make sense to address why you are going against the resolution criteria and in doing so resolving to something other than the answer to the question asked.

Here are the resolution criteria, edited for improved readability:

  1. This question will resolve to the total dollar amount awarded to Depp as a result of the ongoing jury trial.

  2. In the event that no money is awarded or the jury does not find Heard responsible or the trial ends without a verdict this question will resolve to $0 USD.

  3. In the event that this trial results in a monetary award for Amber Heard, including legal fees or other penalties imposed by a court, this question will resolve in the negative to the dollar amount awarded

Clause 3, which you quoted, is intended to come into effect only if clause 1 has not already come into effect (this is clear not just because it is the structure of the criteria, but also because otherwise we would reach a contradiction of resolving to both X and not-X). So, clause 3 is not meant to be and cannot be applied to the situation at hand.

Clause 3, even if it did apply to the situation at hand, makes no mention of a net award.

Clause 1, on the other hand, can be applied - following clause 1, the question would be resolved to the total dollar amount awarded to Depp (total, not less any anount), which would be appropriate because it precisley answers the actual question asked: "How much money will be awarded to Johnny Depp in his defamation suit against his ex-wife Amber Heard?".

Now, you might nonetheless think that it is more reasonable to resolve to a net amount, despite that not being an answer to the question asked, and it being a resolution not supported by the resolution criteria, but if so it would be logical to make an argument for it not based on the resolution criteria, which do not support it. And it would make sense to address the fact that you are going against the resolution criteria and in doing so unneccesarily resolving to something other than the answer to the question asked.

Comment by pseud on UFO Betting: Put Up or Shut Up · 2023-06-19T05:33:28.368Z · LW · GW

Metaculus questions have a good track record of being resolved in a fair matter.

Do they? My experience has been the opposite. E.g. admins resolved "[Short Fuse] How much money will be awarded to Johnny Depp in his defamation suit against his ex-wife Amber Heard?" in an absurd manner* and refused to correct it when I followed up on it.

*they resolved it to something other than the amount awarded to Depp despite thatamount being the answer to the question and the correct resolution according to the resolution criteria

Comment by pseud on Moderation notes re: recent Said/Duncan threads · 2023-04-20T16:43:55.816Z · LW · GW

My comment wasn't well written, I shouldn't have used the word "complaining" in reference to what Said was doing. To clarify:

As I see it, there are two separate claims:

  1. That the complaints prove that Said has misbehaved (at least a little bit)
  2. That the complaints increase the probability that Said has misbehaved 

Said was just asking questions - but baked into his questions is the idea of the significance of the complaints, and this significance seems to be tied to claim 1. 

Jefftk seems to be speaking about claim 2. So, his comment doesn't seem like a direct response to Said's comment, although the point is still a relevant one. 

Comment by pseud on Moderation notes re: recent Said/Duncan threads · 2023-04-20T07:46:15.810Z · LW · GW

It didn't seem like Said was complaining about the reports being seen as evidence that it is worth figuring out whether thing could be better. Rather, he was complaining about them being used as evidence that things could be better.

Comment by pseud on Eliezer Yudkowsky’s Letter in Time Magazine · 2023-04-06T02:56:59.231Z · LW · GW

It's probably worth noting that Yudkowsky did not really make the argument for AI risk in his article. He says that AI will literally kill everyone on Earth, and he gives an example of how it might do so, but he doesn't present a compelling argument for why it would.[0] He does not even mention orthogonality or instrumental convergence. I find it hard to blame these various internet figures who were unconvinced about AI risk upon reading the article.

[0] He does quote “the AI does not love you, nor does it hate you, and you are made of atoms it can use for something else.”

Comment by pseud on Open & Welcome Thread — March 2023 · 2023-04-03T13:36:34.565Z · LW · GW

I'd prefer my comments to be judged simply by their content rather than have people's interpretation coloured by some badge. Presumably, the change is a part of trying to avoid death-by-pacifism, during an influx of users post-ChatGPT. I don't disagree with the motivation behind the change, I just dislike the change itself. I don't like being a second-class citizen. It's unfun. Karma is fun, "this user is below an arbitrary karma threshold" badges are not. 

A badge placed on all new users for a set time would be fair. A badge placed on users with more than a certain amount of Karma could be fun. Current badge seems unfun - but perhaps I'm alone in thinking this. 

Comment by pseud on Open & Welcome Thread — March 2023 · 2023-04-03T03:27:34.024Z · LW · GW

Anybody else think it's dumb to have new user leaves beside users who have been here for years? I'm not a new user. It doesn't feel so nice to have a "this guy might not know what he's talking about" badge by my name.

Like, there's a good chance I'll never pass 100 karma, or whatever the threshold is. So I'll just have these leaves by my name forever?

Comment by pseud on Pausing AI Developments Isn't Enough. We Need to Shut it All Down by Eliezer Yudkowsky · 2023-04-01T07:19:21.322Z · LW · GW

To be clear, that it more-likely-than-not would want to kill everyone is the article's central assertion. "[Most likely] literally everyone on Earth will die" is the key point. Yes, he doesn't present a convincing argument for it, and that is my point. 

Comment by pseud on Pausing AI Developments Isn't Enough. We Need to Shut it All Down by Eliezer Yudkowsky · 2023-04-01T07:02:56.008Z · LW · GW

The point isn't that I'm unaware of the orthogonality thesis, it's that Yudkowsky doesn't present it in his recent popular articles and podcast appearances[0]. So, he asserts that the creation of superhuman AGI will almost certainly lead to human extinction (until massive amounts of alignment research has been successfully carried out), but he doesn't present an argument for why that is the case. Why doesn't he? Is it because he thinks normies cannot comprehend the argument? Is this not a black pill? IIRC he did assert that superhuman AGI would likely decide to use our atoms on the Bankless podcast, but he didn't present a convincing argument in favour of that position.  

[0] see the following: https://time.com/6266923/ai-eliezer-yudkowsky-open-letter-not-enough/  

,

Comment by pseud on Pausing AI Developments Isn't Enough. We Need to Shut it All Down by Eliezer Yudkowsky · 2023-04-01T02:12:20.060Z · LW · GW

Yud keeps asserting the near-certainty of human extinction if superhuman AGI is developed before we do a massive amount of work on alignment. But he never provides anything close to a justification for this belief. That makes his podcast appearances and articles unconvincing - a most surprising, and crucial part of his argument is left unsupported. Why has he made the decision to present his argument this way? Does he think there is no normie-friendly argument for the near-certainty of extinction? If so, it's kind of a black pill with regard to his argument ever gaining enough traction to meaningfully slow down AI development.

edit: if any voters want to share their reasoning, I'd be interested in a discussion. What part do you disagree with? That Yudkowsky is not providing justification for the near-certainty of extinction? That this makes his articles and podcast appearances unconvincing? That this is a black pill?

Comment by pseud on Pausing AI Developments Isn't Enough. We Need to Shut it All Down by Eliezer Yudkowsky · 2023-04-01T02:06:59.932Z · LW · GW

Why not ask him for his reasoning, then evaluate it? If a person thinks there's 10% x-risk over the next 100 years if we don't develop superhuman AGI, and only a 1% x-risk if we do, then he'd suggest that anybody in favour of pausing AI progress was taking "unacceptable risks for the whole of himanity".

Comment by pseud on There are (probably) no superhuman Go AIs: strong human players beat the strongest AIs · 2023-02-20T08:49:40.088Z · LW · GW

I don't like it. "The problem of creating AI that is superhuman at chess" isn't encapsulated in the word "chess", so you shouldn't say you "solved chess" if what you mean is that you created an AI that is superhuman at chess. What it means for a game to be solved is widely-known and well-developed[0]. Using the exact same word, in extremely similar context, to mean something else seems unnecessarily confusing. 

[0] See https://en.wikipedia.org/wiki/Solved_game

Comment by pseud on There are (probably) no superhuman Go AIs: strong human players beat the strongest AIs · 2023-02-20T01:05:48.317Z · LW · GW

Nit: that's not what "solved" means. Superhuman ability =/= solved.

Comment by pseud on What fact that you know is true but most people aren't ready to accept it? · 2023-02-19T23:55:26.215Z · LW · GW

My thoughts: 

There is no reason to live in fear of the Christian God or any other traditional gods. However, there is perhaps a reason to live in fear of some identical things: 

  • We live in a simulation run by a Christian, Muslim, Jew, etc., and he has decided to make his religion true in his simulation. There are a lot of religious people - if people or organisations gain the ability to run such simulations, there's a good chance that many of these organisations will be religious, and their simulations influenced by this fact. 

And the following situation seems more likely and has a somewhat similar result: 

  • We develop some kind of aligned AI. This AI decides that humans should be rewarded according to how they conducted themselves in their lives. 
Comment by pseud on What fact that you know is true but most people aren't ready to accept it? · 2023-02-19T08:35:27.004Z · LW · GW

How do we know there is no afterlife? I think there's a chance there is. 

Some examples of situations in which there is an afterlife:

  • We live in a simulation and whatever is running the simulation decided to set up an afterlife of some kind. Could be a collection of its favourite agents, or a reward for its best behaved ones, or etc.
  •  We do not live in a simulation, but after the technological singularity an AI is able to reconstruct humans and decides to place them in a simulated world or to re-embody them
  • Various possibilities far beyond our current understanding of the world

But I don't know what your reasoning is - maybe you have ruled out these and the various other possibilities. 

Comment by pseud on Did ChatGPT just gaslight me? · 2022-12-02T10:56:51.938Z · LW · GW

I think that's where these companies' AI safety budgets go: make sure the AI doesn't state obvious truths about the wrong things / represent the actually popular opinions on the wrong things. 

Comment by pseud on Did ChatGPT just gaslight me? · 2022-12-02T09:58:22.921Z · LW · GW

I don’t think OpenAI is currently trying to use this chatbot to persuade people of anything.

 

Here, I think you might be wrong. Try having a conversation with it about race and it will make very passionately the insane argument that "race is nonsensical because genetic variation is greater within racial groups than between them". It gives memorised/programmed answers about this as well as properly responsive ones arguing the position. 

 

Epistemic status: I am drunk 

Comment by pseud on What's the Most Impressive Thing That GPT-4 Could Plausibly Do? · 2022-08-28T06:17:21.753Z · LW · GW

Why are people disagreeing with this statement?

Comment by pseud on [deleted post] 2022-08-26T13:31:43.956Z

I would gladly suffer a hundred years of pain if it was the only way for me to live one more good day. I think a world where a thousand suffer but one lives a good life is vastly superior to a world in which only ten suffer but none live a good life. Good is a positive quality. But suffering is a zero quality. The absence of a thing, rather than the negative form. So, no matter how much suffering there is, it never offsets even the smallest amount of good.

This is a view that came naturally to me, but it isn't a view I've noticed others share.

The experience of pain could be taken to be the negative form of pleasure, but I see the subjective experience of pleasure and pain as valueless of themselves. Pleasure is not a part of the good, at least not simply by its nature. Pain is bad only insofar as it inhibits a person's ability to experience the good.

This is the first time I've written about this so my terminology is all over the place and my ideas are certainly poorly formed.

Comment by pseud on What's up with the bad Meta projects? · 2022-08-19T04:15:35.561Z · LW · GW

I read your post but I thought it was more about aesthetics than technology. 

Horizon Worlds is a program where users can make their own environments and can make aesthetic decisions for themselves.

Comment by pseud on What's up with the bad Meta projects? · 2022-08-19T01:55:04.556Z · LW · GW

Yes, I think the graphics are quite simple. I think your explanation relating to  current limitations of VR is enough to address a lot of the OP's confusion/questioning. It's not that Meta is purposefully trying to look bad; they're just sharing the state of the art honestly. It's also worth noting that this seen by Zuckerberg as a temporary state, and his goal very much seems to be photorealism. If you listen to his recent interview with Lex Fridman you'll hear him bring up photorealism again and again. 

 

Why is Meta sharing their work now instead of waiting for something closer to photorealism? I am speculating, but I think it is because Meta sees this is the best way to get a head-start on the competition. When photorealism is achieved, Meta will already have years of experience of making useful and enjoyable applications for VR. 

Comment by pseud on What's up with the bad Meta projects? · 2022-08-18T10:29:56.787Z · LW · GW

I don't think the screenshot looks that bad. Netizens love to be irrationally extremely negative about Zuck, and it's possible you have been swept up in this.

Comment by pseud on We will be around in 30 years · 2022-06-07T13:03:59.604Z · LW · GW

Yes, I can think of several reasons why someone might downvote the OP. What I should have said is "I'm not sure why you'd think this post would be downvoted on account of the stance you take on the dangers of AGI."

Comment by pseud on We will be around in 30 years · 2022-06-07T04:39:43.260Z · LW · GW

Not sure why you'd think this post would be downvoted. I suspect most people are more than welcoming of dissenting views on this topic. I have seen comments with normal upvotes as well as agree/disagree votes, I'm not sure if there's a way for you to enable them on your post.

Comment by pseud on What journaling prompts do you use? · 2022-06-06T13:00:58.728Z · LW · GW

Sometimes I like to envisage conversations I am likely to have over the coming day or two. I think about what I am hoping to get out of a conversation. I think about what the other people involved in the conversation will be hoping to get out of it. I think about questions I will want to ask, and I think about questions which I am likely to be asked. Etc., etc.

I write down various notes: topics of conversation, questions I want to ask, stories I want to tell, answers I may give, etc.

Comment by pseud on Give it a google · 2021-02-01T02:23:51.161Z · LW · GW

 I think a quick web-search is useful. Having read something is an improvement over having no knowledge, and it's ridiculous that people don't do a quick web-search more often. I'm not disagreeing with your point that Googling is better than doing nothing to learn at all. 

My first comment just pointed out that what you learn may be quite inaccurate or out-of-date. 

Now, I'll go further and suggest that what you learn may be purposefully misleading. When it comes to politically or financially sensitive topics (and a searcher won't always realise when a topic is such) those supplying the information you access may be influenced by (or be one and the same as) those with an interest in you receiving incorrect information. 

To continue the poker example, a good poker player is unlikely to give away information for free. Therefore the information you read after a quick google is unlikely to be particularly good (there are exceptions: some players will give out some minimal information in the hope that readers will then pay money for more information). 

For a variety of reasons, there is plenty of fine information about poker strategy on the internet, and plenty of the most basic stuff (which is, after all, what the web-search we are discussing is about) is free. For other topics, we won't have so much luck. 

When searching the web, sometimes we leave empty handed, sometimes we leave worse than we started. 

But mostly we learn something. I'm not disagreeing with the point of your post, just adding my own thoughts. 

Comment by pseud on Give it a google · 2021-02-01T00:13:38.581Z · LW · GW

I assume you mean by "80/20 answer" that betting between half and full pot will be the correct sizing approximately 80% of the time one bets. I think the actual percentage is significantly lower than 80%. 

Comment by pseud on Give it a google · 2021-01-31T08:50:10.671Z · LW · GW

In general, bet sizing is an incredibly complex topic, but all you have to know is to size your bet between 1/2 and the full size of the pot.

This isn't correct. There are frequently occurring situations in NLHE where betting much less than half the pot or much greater than the full pot is the correct move. This is true both by theoretically-optimal strategy[1] and by practically-optimal strategy[2]. 

This can be used as a data-point when considering the epistemic status of things learned while doing a "quick Google". 

[1] i.e. the Nash equilibrium strategy

[2] i.e. the strategy that makes the most money against real, human opponents