Replacing Karma with Good Heart Tokens (Worth $1!)

post by Ben Pace (Benito), habryka (habryka4) · 2022-04-01T09:31:34.332Z · LW · GW · 173 comments

Contents

    Voting, Leaderboards and Payment Info
  Why are we doing this?
    The EA Funding Ecosystem Wants To Fund Megaprojects
    Trends Generally Continue Trending
    Yes, We Have Taken User Feedback (GPT-10 Simulated)
      A Final Note on Why
    Future Enhancement with Machine Learning
  Good Hearts Laws
None
183 comments

Starting today, we're replacing karma with Good Heart Tokens which can be exchanged for 1 USD each.

We've been thinking very creatively about metrics of things we care about, and we've discovered that karma is highly correlated with value

Therefore, we're creating a token that quantifies the goodness of the people writing, and whether in their hearts they care about rationality and saving the world.

We're calling these new tokens Good Heart Tokens. And in partnership with our EA funders, we'll be paying users $1 for each token that they earn.

"The essence of any religion is a good heart [token]." 
— The Dalai Lama[1]

Voting, Leaderboards and Payment Info

Comments and posts now show you how many Good Heart Tokens they have.

(This solely applies to all new content on the site.)

At the top of LessWrong, there is now a leaderboard to show the measurement of who has the Goodest Heart. It looks like this. (No, self-votes are not counted!)

The usernames of our Goodest Hearts will be given a colorful flair throughout the entirety of their posts and comments on LessWrong.

To receive your funds, please log in and enter your payment info at lesswrong.com/payments/account [? · GW].

While the form suggests using a PayPal address, you may also add an Ethereum address, or the name of a charity that you'd like us to donate it to.

Why are we doing this?

On this very day last year, we were in a dire spot [LW · GW].

We really, really needed the money.

To fund our ever-increasing costs, we were forced to move to Substack and monetize most of our content.

Several generous users subscribed at the price of 1 BTC/month, for which we will always be grateful. It turns out that Bitcoin was valued a little higher than the $13.2 we had assumed [LW · GW], and this funding quickly allowed us to return the site to its previous state.

Once we restored the site, we still had a huge pile of money, and we've spent the last year desperately trying to get rid of it.

In our intellectual circles, Robin Hanson has suggested making challenge coins, and Paul Christiano has suggested making impact certificates [EA · GW]. Both are tokens that can later be exchanged for money, and whose value correlates with something we care about.

Inspired by that, we finally cracked it, and this is our plan.

...We're also hoping that this is an initial prototype that larger EA funders will jump on board to scale up!

The EA Funding Ecosystem Wants To Fund Megaprojects

"A good heart [token] is worth gold."
— King Henry IV, William Shakespeare[2]

Effective altruism has always been core to our hearts, and this is our big step to fully bring to bear the principles of effective altruism on making LessWrong great.

The new FTX Future Fund has said:

We’re interested in directly funding blogs, Substacks, or channels on YouTube, TikTok, Instagram, Twitter, etc.

They've also said:

We’re particularly interested in funding massively scalable projects: projects that could scale up to productively spend tens or hundreds of millions of dollars per year.

We are the best of both worlds: A blog that FTX and other funders can just pour money into. Right now we're trading $1 per Good Heart Token, but in the future we could 10x or 100x this number and possibly see linear returns in quality content!

Paul Christiano has said [LW · GW]:

I have a general read of history where trend extrapolation works extraordinarily well relative to other kinds of forecasting, to the extent that the best first-pass heuristic for whether a prediction is likely to be accurate is whether it's a trend extrapolation and how far in the future it is.

We agree with this position. So here is our trend-extrapolation argument, which we think has been true for many years and so will continue to be true for at least a few years.

The blue line is the observed trend, the yellow dotted line is the extrapolation.

So far it seems like higher-karma posts have been responsible for better insights about rationality and existential risk. The natural extrapolation suggests it will increase if people produce more content that gets high karma scores. Other trends are possible, but they're not probable. The prior should be that the trend continues!

However, epistemic modesty does compel me to take into account the possibility that we are wrong on this simple trend extrapolation. To remove any remaining likelihood of karma and intellectual progress becoming decoupled, we would like to ask all users participating in the Good Hearts Project to really try hard to not be swayed by any unaligned incentives (e.g. the desire to give your friends money).

Yes, We Have Taken User Feedback (GPT-10 Simulated)

We care about hearing arguments for and against our decision, so we consulted with a proprietary beta version of GPT-10 to infer what several LessWrong users would say about the potential downsides of this project. Here are some of the responses.

Eliezer Yudkowsky [LW · GW]: "To see your own creation have its soul turned into a monster before your eyes is a curious experience."

Anna Salamon [LW · GW]: "I can imagine a world where earnest and honest young people learn what's rewarded in this community is the most pointed nitpick possible under a post and that this might be a key factor in our inability to coordinate on preventing existential risk".

Kaj Sotala [LW · GW]: "I am worried it would lead to the total collapse of an ecosystem I've poured my heart and soul into in the past decade".

Scott Garrabrant [LW · GW]: "Now I can get more Good Heart tokens than Abram Demski! This is exactly what I come to LessWrong for."

Clippy [LW · GW]: "What's the exchange rate between these tokens and paperclips?"

Quirinus Quirrell [LW · GW]: "A thousand of your measly tokens pale in comparison to a single Quirrell point."

Zvi Mowshowitz [LW · GW]: "I would never sell my soul for such a low price. LessWrong delenda est." (Then my computer burst into flame, which is why this post was two hours late.)

(Of course, this is only what our beta-GPT-10 simulations of these users said. The users are welcome to give their actual replies in the comment section below.)

O​ur EA funders have reviewed these concerns, and agree that there are risks, but think that, in the absence of anything better to do with money, it's worth trying to scale this system.

A Final Note on Why

I think the work on LessWrong matters a lot, and I'd like to see a world where these people and other people like them can devote themselves full-time to producing such work, and be financially supported when doing so.

Future Enhancement with Machine Learning

We're hoping to enhance this in the future by using machine learning on users' content to predict the karma you will get. Right now you only get Good Heart tokens after your content has been voted on, but with the right training, we expect our systems will be able to predict how many tokens you'll receive in advance.

This will initially look like people getting the money for their post at the moment of publishing. Then it will look like people getting the money when they've opened their draft and entered the title. Eventually, we hope to start paying people the moment they create their accounts. 

For example, Terence Tao will create a LessWrong account, receive $10 million within seconds, and immediately retire from academia.

Good Hearts Laws

"Together, people with good hearts [tokens] and fine minds can move the world."  
— Bill Clinton[3]

  1. ^
  2. ^
  3. ^

173 comments

Comments sorted by top scores.

comment by Vaniver · 2022-04-01T16:22:18.386Z · LW(p) · GW(p)

Why have "notice your surprise [LW · GW]" day when you could instead have "publish all of your drafts" day? ;)

Replies from: lsusr
comment by lsusr · 2022-04-01T23:37:16.531Z · LW(p) · GW(p)

I have so many unpublished drafts that I hit the limit (5) of what LW allows in a day. When I hit the limit, I took the least popular one down and put up a better post instead.

Replies from: hath
comment by hath · 2022-04-02T02:48:00.851Z · LW(p) · GW(p)

If you didn't already try, I bet Lightcone would let you post more if you asked over Intercom.

Replies from: Vaniver
comment by Vaniver · 2022-04-02T03:28:52.994Z · LW(p) · GW(p)

I'm guessing it would just take us editing the relevant posts, and so would be technically easy; I think it might be a bad idea to do it at this point, since there wouldn't be that long for the posts to be read (and it'd dilute attention between them).

comment by habryka (habryka4) · 2022-04-01T20:59:15.446Z · LW(p) · GW(p)

With deep surprise and betrayal and regret in my heart am I forced to announce that we noticed highly suspicious voting and posting activity from three suspicious user engaging in a criminal conspiracy:

We reported this case to the Good Hearts police, who have seized all Good Heart Tokens from the three individuals, and fined them each 1000 Good Heart Tokens.

While we never thought it would come to this, we are clarifying that voting rings are not allowed, and will be met with harsh punishment.

Replies from: aphyer, johnswentworth, localdeity
comment by aphyer · 2022-04-01T23:18:19.355Z · LW(p) · GW(p)

My client acknowledges his guilt but wishes to appeal the harshness of the punishment.

My client has renounced his evil ways, and sworn to follow the straight and narrow path.  As proof of his sincere remorse, he has submitted two new articles (1 [LW · GW], 2 [LW · GW]), intending to pursue Good Hearts with honestly-written articles earning honest upvotes.

However, the severity of the fine has made it impossible for him to ever return to a normal life in society.  The GH1000 penalty is as of this writing ~5 times the largest quantity of GH earned by any writer.  

My client is additionally willing to delete the criminal posts, just as soon as someone tells him how to delete a draft (it's unclear to me how you can actually do this).

In the interests of clemency and rehabilitation, Your Honor, I ask that you reduce my client's penalty to a sum he can more realistically 'work off' through good deeds, rather than fining him an amount far in excess of what even the wealthiest GH-holders are able to pay.

Replies from: Raemon
comment by Raemon · 2022-04-02T03:35:39.191Z · LW(p) · GW(p)

After much careful evaluation of this counterargument, examining the value of your posts, wringing my hands, and then kinda winging it because I didn't feel like spending that much effort carefully counting up all the relevant bits of karma, I decided to set your karma to -22, and johnswentworth's to 114 (each of these are the values of your top-level posts, minus 50, which seemed like a reasonable penalty)

Replies from: aphyer
comment by aphyer · 2022-04-02T03:48:54.747Z · LW(p) · GW(p)

Thank you!

comment by johnswentworth · 2022-04-01T22:37:48.339Z · LW(p) · GW(p)

This isn't the last you'll see of us, Good-Hearter! You may have won a token victory, but we'll be back. Muahahahaha!

comment by localdeity · 2022-04-01T21:43:13.502Z · LW(p) · GW(p)

I am shocked, shocked, to find voting rings in this forum!

Your upvotes, sir.

Ah, thank you very much.  Everybody out at once!

comment by abramdemski · 2022-04-01T16:31:03.152Z · LW(p) · GW(p)

Strong-downvoted to deprive Ben Pace of money, because mwahaha.

[After thinking about this more, I changed my mind and upvoted]

Question: where can I see my Good Hearts score if I'm not currently on the leaderboard?

Assertion: lsusr appears to be setting a good example by engaging with this in good faith, posting lots of actually good stuff today. lsusr is also currently in the lead!!! 

This provides actual evidence that this is actually an actual good idea (at least, a good idea for an April 1st one-shot)

If you are ALSO engaging with this in good faith, comment here to let me know. This will reduce the chance that I miss the good stuff you post today. (IE, I'll consider upvoting it in good faith.)

Replies from: Vaniver, hath, Keller, lsusr
comment by Vaniver · 2022-04-01T16:32:42.936Z · LW(p) · GW(p)

I'm engaging with this in good faith, tho I think my current effortpost-draft is one that I want reviewed by other people before I post it, which suggests either I should switch to a different idea that I can shoot from the hip with or just do comment-engagement instead of post-engagement today.

 

[edit] Also I think we should just keep the leaderboard? I remember being pretty motivated by the 'top karma last 30 days' leaderboard to post something about once a month.

comment by hath · 2022-04-01T16:44:17.282Z · LW(p) · GW(p)

I didn't realize that this was a test-bed for seeing if monetary value would actually work; everything I post on this account from now on will be actual effort posts that I would post normally. I really don't want to be the type of person who sees the Petrov Day button and automatically tries a bunch of random passwords, in the process ruining it for everyone. I also think that monetary rewards for LW posts might be a good idea, and want to help operating by the decision-theory of whatever would work well in an actual LW environment with financial rewards; this means passing up the (12 eligible comments + however many other comments I make)*2 dollars which would result from me strong upvoting every comment w/ the test account I made yesterday to try and debug something. (posting this is also me committing myself to NOT doing that).

comment by Keller · 2022-04-03T15:17:35.421Z · LW(p) · GW(p)

I have made a good-faith post as a result of the extension!

comment by lsusr · 2022-04-01T23:51:33.949Z · LW(p) · GW(p)

You can view your Good Heart Token count on your account page [LW · GW] next to your karma.

comment by Kaj_Sotala · 2022-04-01T12:50:55.062Z · LW(p) · GW(p)

Not sure why your simulation of me didn't point this out, but it seems odd to call something a "token" and then not make it into a cryptocurrency. Maybe this was the original plan, but then you realized that launching a GoodHeartCoin would make you even richer, thus worsening the very problem you are trying to solve?

Replies from: Benito
comment by Ben Pace (Benito) · 2022-04-04T04:55:46.310Z · LW(p) · GW(p)

It was just so in an earlier draft. 

I was debating between LessWrongCoin versus the more succinct WrongCoin.

comment by mingyuan · 2022-04-01T13:39:18.031Z · LW(p) · GW(p)

You can't just cite a Shakespeare quote as being from King Henry IV, that is two different plays you BARBARIAN. If you were trying to pander to me because King Henry IV Part 1 is my favorite play, you FAILED; you quoted from King Henry IV Part 2. Disgusting.

comment by Alex_Altair · 2022-04-01T14:48:16.216Z · LW(p) · GW(p)

Next thing you know they'll be announcing on the Alignment Forum that they've founded the Ally Mint, a foundation for minting tokens to award to our allies in alignment research.

Replies from: Pattern
comment by Pattern · 2022-04-03T06:50:12.935Z · LW(p) · GW(p)

Allies? Just take out the middleman, and pledge to give the tokens to AIs for being good.

Replies from: donald-hobson, Benito
comment by Donald Hobson (donald-hobson) · 2022-04-07T16:20:08.187Z · LW(p) · GW(p)

Ah, but you can't measure whether or not the AI is doing something that is actually good or just something that seems good to you. That is part of the problem.

Also it isn't quite clear why a super-intelligence would want some not very valuable crypto tokens more than it would want the atoms that used to make up the economic system that made the crypto valuable.

Replies from: Pattern
comment by Pattern · 2022-04-08T18:24:33.285Z · LW(p) · GW(p)
Ah, but you can't measure whether or not the AI is doing something that is actually good or just something that seems good to you. That is part of the problem.

This is also a problem for the

minting tokens to award to our allies in alignment research

idea.


I was joking that, crypto[currency] is about taking out the third parties, so the same should be done there.

comment by Ben Pace (Benito) · 2022-04-03T06:54:17.843Z · LW(p) · GW(p)

But Good Mint isn't a pun.

Replies from: Pattern
comment by Pattern · 2022-04-03T18:22:20.372Z · LW(p) · GW(p)

Ally, aligned. Close enough.

comment by Alex_Altair · 2022-04-01T17:41:51.687Z · LW(p) · GW(p)

I'm curious what people think the natural exchange rate is between dollars and LW karma. Presumably it's not zero, although I think it's definitely less than $1; it takes me a lot more than $100 worth of effort to write a 100 karma post.

It's a little bit weird to think about "how much would I pay for karma", because the whole point of karma is that you can only get it through creating a specific kind of value, and if anyone could buy it then the whole system would lose it's purpose. But it feels more natural to ask "how many karma is an hour of my time worth", and then I can convert that through "how many dollars is an hour of my time worth".

Replies from: Lanrian, abramdemski, lsusr, AprilSR, Pattern
comment by Lukas Finnveden (Lanrian) · 2022-04-01T17:51:17.770Z · LW(p) · GW(p)

Unfortunately, it's generally a lot easier to generate karma through commenting than through posting.

Once upon a time, I hear there was a 10x multiplier on post karma. 10x is a lot, but it seems pretty plausible to me that a ~3x multiplier on post karma would be good.

Replies from: tomcatfish, istandleet
comment by Robert Kennedy (istandleet) · 2022-04-01T22:09:04.888Z · LW(p) · GW(p)

I think top level posts generate much more than 10x the value than the entire comments section combined, based off my impression that the majority of lurkers don't get deep in the comments. I wonder if top level posts having a x^1.5 exponent would get closer to the ideal... That would also disincentivize post series...

comment by abramdemski · 2022-04-01T18:07:23.643Z · LW(p) · GW(p)

You could instead ask: how much would you be willing to pay for 1 karma's worth of the-sort-of-value-karma-measures?

Or perhaps: how much would the LW community altogether be willing to pay?

We could estimate the second by dividing the amount of money Scott Alexander makes on substack per month (where do I find this?) by our estimate of the monthly LW karma those articles would generate, if they appeared on LW. (100 per post??)

Replies from: lsusr
comment by lsusr · 2022-04-01T23:58:07.760Z · LW(p) · GW(p)

The value of a post grows faster than its karma. A 200 karma post is more than 2× harder to write and provides more than 2× the value of a 100 karma post. A 100 karma post is more than 2× harder to write and provides more than 2× the value of a 50 karma post.

comment by lsusr · 2022-04-01T23:55:52.723Z · LW(p) · GW(p)

The LW team has, in the past, offered $500 for quality book reviews [LW · GW] with a minimum post karma of about 50.

comment by AprilSR · 2022-04-02T00:27:27.017Z · LW(p) · GW(p)

I could probably be paid to write medium-to-decent quality LW posts pretty cheaply? I think today has given me an affordance for "write things for LW" that I didn't use to have nearly as strongly.

comment by Pattern · 2022-04-08T18:26:50.191Z · LW(p) · GW(p)

How much would you be willing to pay someone to write a 100(+) karma post?

comment by hath · 2022-04-01T16:28:25.923Z · LW(p) · GW(p)

To remove any remaining likelihood of karma and intellectual progress becoming decoupled, we would like to ask all users participating in the Good Hearts Project to really try hard to not be swayed by any unaligned incentives (e.g. the desire to give your friends money)

...I actually didn't see this until now. Huh.

Replies from: hath
comment by hath · 2022-04-01T16:31:02.903Z · LW(p) · GW(p)

I guess this means it's actually a test we're supposed to pass. I have a feeling that there's going to be some confusion over "what game are we really playing", like with Petrov Day. Oh. It's to see if we could actually, long-term, reward LW participation with money. Guess I'll just focus on actually writing posts.

Man, this is the worst time for me to have only skimmed one of those posts.

comment by hath · 2022-04-01T19:51:08.728Z · LW(p) · GW(p)

It's really interesting seeing the change in attitude toward low-effort asking-for-money posts. Earlier, people upvoted/put up with them; now people are actively punishing bullshit with strong downvotes. This is good for LW implementing monetary incentives in the future; we can punish Goodharters ourselves.

comment by PhilGoetz · 2022-04-01T17:57:35.573Z · LW(p) · GW(p)

Finding a way for people to make money by posting good ideas is a great idea.

Saying that it should be based on the goodness of the people and how much they care is a terrible idea.  Privileging goodness and caring over reason is the most well-trodden path to unreason.  This is LessWrong.  I go to fimfiction for rainbows and unicorns.

Replies from: hath
comment by hath · 2022-04-01T18:32:43.824Z · LW(p) · GW(p)

I think that was part of the whole "haha goodhart's law doesn't exist, making value is really easy" joke. However, it's also possible that that's... actually one of the hard-to-fake things they're looking for (along with actual competence/intelligence). See PG's Mean People Fail or Earnestness. I agree that "just give good money to good people" is a terrible idea, but there's a steelman of that which is "along with intelligence, originality, and domain expertise, being a Good Person (whatever that means) and being earnest is a really good trait in EA/LW and the world at large, and so we should try and find people who are Good and Earnest, to whatever extent that we can make sure that isn't Goodharted ."

(I somewhat expect someone at LW to respond to this saying "no, the whole goodness thing was a joke")

comment by lc · 2022-04-01T13:50:16.091Z · LW(p) · GW(p)

Literally took me until the end of the post to figure out the pun.

Replies from: Simon Skade, Bucky
comment by Towards_Keeperhood (Simon Skade) · 2022-04-01T22:16:47.210Z · LW(p) · GW(p)

For people like me who are really slow on the uptake in things like this, and realize the pun randomly a few hours later while doing something else: The pun is because of goodhart (from Goodhart's law).) (I'm not thinking much in what a word sounds like, and I just overread the "Good Hearts Laws" as something not particularly interesting, so I guess this is why I haven't noticed.)

comment by Bucky · 2022-04-01T19:48:07.854Z · LW(p) · GW(p)

That makes me feel less bad for doing the same...

comment by AprilSR · 2022-04-01T15:25:31.285Z · LW(p) · GW(p)

Can I request a 2-dimensional version of this? What if I want to distinguish between good comments and comments I agree with?

Replies from: Viliam, abramdemski
comment by Viliam · 2022-04-01T18:40:44.810Z · LW(p) · GW(p)

Hearts for good comments (monetary value $1), likes for the comments I agree with (monetary value $0). Because we obviously want to incentivize good content rather that political factions.

The advantage is that this will be backwards compatible with the old karma system, where people upvoted comments they agreed with, and received $0 for such comments.

comment by abramdemski · 2022-04-01T17:59:36.440Z · LW(p) · GW(p)

Good comments should net real dollars, while agreed-with comments net imaginary dollars (that is, a currency which equals the square root of real-dollar debt).

comment by AprilSR · 2022-04-01T15:04:07.665Z · LW(p) · GW(p)

If I get negative karma do I need to pay LessWrong money

Replies from: abramdemski
comment by abramdemski · 2022-04-01T15:47:27.305Z · LW(p) · GW(p)

Think of it as a LessWrong fundraiser.

comment by Measure · 2022-04-01T11:49:31.685Z · LW(p) · GW(p)

I'm worried that if I exchange my Good Heart Tokens for sordid monetary profits, this will reduce the goodness of my heart. Do the payments replace the Tokens that earn them, or does my Heart remain Good?

Replies from: hath
comment by hath · 2022-04-01T16:24:12.505Z · LW(p) · GW(p)

Next April Fools, Lightcone will allow remaining Good Heart Tokens to be exchanged for longer-lifespan; when AGI finally gets around to making us grey goo, each token spent will buy a nanosecond of extra life.

Also, you can't measure your Heart's purity, so there's no point trying to protect that value. Simple utility functions are so much easier to carry out. For guidance, look at the everyday rock; no messy values like "lack of boredom" or "happiness".

Replies from: Pattern
comment by Pattern · 2022-04-03T06:52:22.286Z · LW(p) · GW(p)

Just exchange the money for covid tests.

comment by Tornus · 2022-04-01T13:46:54.231Z · LW(p) · GW(p)

This is a really good start, and I look forward to the inevitable improvements in the quality of discourse. But to fully leverage the potential of this exciting new system, I think you should create a futures market so we can bet on (or against) specific individuals writing good posts in future.

Also: from now on, I vow to ignore any and all ideas that aren't supported by next-level puns.

comment by Yoav Ravid · 2022-04-01T10:12:52.112Z · LW(p) · GW(p)

Brilliant, and a marvelous pun.

comment by aphyer · 2022-04-01T20:45:43.069Z · LW(p) · GW(p)

I appear to have disappeared from the leaderboard despite my large volumes of high-quality content and justly earned Good Hearts. Is there a display bug of some kind?

Replies from: MondSemmel
comment by MondSemmel · 2022-04-01T21:05:17.008Z · LW(p) · GW(p)

Apparently you ran afoul of the Good Hearts police [LW(p) · GW(p)] :p.

comment by Gordon Seidoh Worley (gworley) · 2022-04-01T19:52:21.713Z · LW(p) · GW(p)

I've updated the donation lottery [LW · GW] to reflect the update [LW · GW] to the program. Thus I'm canceling this bit of fun since it doesn't make sense after April Fools Day.

So I'm attempting to run a donation lottery [LW · GW] off the new system, which is fun. However, it'd also be nice to end up with the Goodest Heart.

I've already promised on that post to subsidize the donation lottery up to the $600 level if it turns out there's actually no money (or I just get less than $600 worth of votes).

However, in the interest of getting the most internet points I can today, if at the end of the Good Heart Token experiment I end up with the Goodest Heart (I'm at the top of the leaderboard) I'll donate an additional $1000 to the winner of the donation lottery.

I'll make a similar comment over on the donation lottery.

To ensure I actually get this, if someone else makes a competing offer to try to get my votes (for example, promises to donate $1001 to the winner of the donation lottery if they end up with the Goodest Heart, or otherwise makes a bid for votes to have the Goodest Heart) and they do in fact end up with the Goodest Heart, I'll instead donate $1000 to whatever charity I think the person who usurps me would most dislike to see a donation made to (that is not a charity that is net-negative for the world in my estimation).

May the voting commence!

comment by JenniferRM · 2022-04-01T17:07:46.015Z · LW(p) · GW(p)

I'm so glad that comments with pointed nitpicking of world saving ideas are getting the respect and remuneration they deserves!

That said... I think maybe you've never heard of quadratic voting, which is VERY TRENDY lately?

Based on reasoning that I can unpack if you're not smart enough to steelman my arguments and understand it for yourself, you should consider giving mere OPs the natural log of their upvotes, while giving e^upvotes good hearts for votes on comments.

The deep logic here is based on the insight that the hardest part of writing is figuring out who your audience is, and what they don't understand, and whether they would welcome what you're writing and be helped by it.

This is why brutally nitpicking writing by authors from a different social bubble than you is such an important part of a healthy epistemic ecology... if potential authors didn't have examples of brutal nitpicking here on Lesswrong, they wouldn't be able to understand and imagine the very high quality of the audience around these parts. 

Also, the bad posters wouldn't be pre-emptively scared away from writing dumb things. 

And most importantly, the good posters wouldn't write such consistently good posts.

We know that GANs are the heart of most important advances in machine learning. Smart people can't help but notice that comments have that same role here on Lesswrong.

Therefore, today of all days, on the anniversary of this wonderful program, I think it is important (A) to remind people to think very hard, and then (B) to praise people who have the strongest pruning skills [LW · GW] and least disinclination to verbosely criticize online acquaintances in public, and then (C) to affirm that the good heart system should be tweaked (again) to make sure that "critics in the comments" get the most money per upvote, since comment upvotes are some of the most causally significant upvotes that can be generated by the discerning, wise, rational, and not-at-all-emotionally-manipulable voters of Lesswrong <3

Replies from: Vaniver
comment by Vaniver · 2022-04-01T17:48:31.105Z · LW(p) · GW(p)

That said... I think maybe you've never heard of quadratic voting, which is VERY TRENDY lately?


Note that quadratic voting has been used for all of the annual reviews, starting in 2018 [LW · GW].

Replies from: Kaj_Sotala
comment by Kaj_Sotala · 2022-04-01T18:33:13.683Z · LW(p) · GW(p)

I assumed that that was the joke.

comment by P. · 2022-04-01T10:02:50.456Z · LW(p) · GW(p)

Given the flood of comments that will inevitably result from this, it might be hard to get noticed and to surface the best ones to the top. So I am offering the following service: If you reply to this I guarantee that I will read your comment, and then will give you one or two upvotes (or none) depending on how insightful I consider it to be. Sadly, this only works if people get to see this comment, so it is in your best interest to upvote it. Let’s turn this into a new, better comment section!

Replies from: abramdemski, Simon Skade
comment by abramdemski · 2022-04-01T17:52:23.468Z · LW(p) · GW(p)

This was very clever, so I strong-upvoted it, but it also feels like it violates good faith, so I then strong-downvoted it.

comment by Towards_Keeperhood (Simon Skade) · 2022-04-01T14:43:12.272Z · LW(p) · GW(p)

Agreed, since now many people will probably comment in this thread, I make the same recursive offer:

If you reply to this I guarantee that I will read your comment, and then will give you one or two upvotes (or none) depending on how insightful I consider it to be.

So please upvote this comment so it stays on top of this comment thread!

Replies from: P.
comment by P. · 2022-04-01T16:22:23.675Z · LW(p) · GW(p)

I upvoted you (+1), but the point of my comment was to make sure that someone besides the author of a comment reads it so that the score more closely reflects its real value. If you do it recursively then we will have two kinds of comments to consider: replies to my comment and to your, so it will become harder to make comparisons.

I guess I should have said something like “and if you up/downvote a reply in this comment thread, please read and score all of its siblings accordingly”, since if it's just me upvoting the problem will persist. I can only give one of 5 possible scores, they will be just a noisy estimate of the “ideal score” and older comments will have bigger magnitudes.

And even if that was fixed, it seems people are mostly ignoring what I said and commenting elsewhere. Oh well, at least I tried.

When I first wrote the comment, it seemed like the best possible strategy according to what I care about. It goodharts the LW karma system so that it loses correlation with intellectual progress or whatever else it is supposed to reflect. And yet it isn’t a pure “I will unfairly upvote you and you unfairly upvote me” like aphyer’s and Tofly’s comments, nor is it nonsense or bot-like behaviour, like what rank-biseral and maybe goodestheart are doing. Now I’m not so sure.

To the Lightcone team, if you want to make sure that scores actually reflect the quality of comments and that users are shown the best possible ones, you should probably implement a multi-armed bandit algorithm as the default sorting method, divide the scores by the number of impressions and then multiply by some constant (10 times the standard deviation?) so that they aren’t tiny. For comments with very few votes you could use some ML model as a prior or show the uncertainty to the user.

Replies from: P.
comment by P. · 2022-04-01T16:45:07.818Z · LW(p) · GW(p)

Oh, and another bad faith strategy consists on writing many ok comments instead of just a few good ones, the Good Heart Tokens a user receives aren’t actually proportional to the total value added. Please keep the score of this one at 1.

comment by philip_b (crabman) · 2022-04-01T09:21:27.085Z · LW(p) · GW(p)

There is an inconsistency in the formatting of (simulated) user feedback. Some are formatted as Username: "XXX", e.g.

Anna Salamon: "I can imagine a world where earnest and honest young people learn what's rewarded in this community is the most pointed nitpick possible under a post and that this might be a key factor in our inability to coordinate on preventing existential risk".

while others are formatted as Username said "XXX", e.g.

Eliezer Yudkowsky said "To see your own creation have its soul turned into a monster before your eyes is a curious experience."

Replies from: Benito
comment by Ben Pace (Benito) · 2022-04-01T09:31:42.944Z · LW(p) · GW(p)

you saw nothing

comment by GabrielExists (gabriel-1) · 2022-04-01T16:53:16.705Z · LW(p) · GW(p)

4 hours ago, I thought there was a 0% chance money would be paid out. But 3 hours ago, a discussion made me think there was a chance they would do a payout for today, and then retract the whole thing tomorrow. That made me 20% certain that money would be paid out. So between those hours, there was a linear increase in the chance of money being paid out. And as other trends are possible, but not probable, then the prior should be that the trend continues, right? So I expect this to arrive at 100% chance several hours before payout time.

Though I'm curious whether the community agrees:

https://manifold.markets/LeonardoTaglialegne/will-lesswrong-actually-pay-out-the 

Replies from: Alex_Altair, rossry
comment by Alex_Altair · 2022-04-01T17:46:10.591Z · LW(p) · GW(p)

This part makes me think they'll actually pay out;

  • There is currently a cap of 600 Good Heart Tokens per user while we are rolling out the system.
  • The minimum number of tokens to be exchanged is 25.

Especially the part where it's 600 and not 500. That sounds like the kind of number that came out of people actually deciding how much they wanted to pay out.

Replies from: Vaniver
comment by Vaniver · 2022-04-01T17:49:56.709Z · LW(p) · GW(p)

Especially the part where it's 600 and not 500. That sounds like the kind of number that came out of people actually deciding how much they wanted to pay out.

It's the kind of number caused by government regulation.

comment by rossry · 2022-04-01T16:59:45.295Z · LW(p) · GW(p)

Not taking extrapolation far enough!

4 hours ago, your expected value of a point was $0. In an hour, it increased to $0.2, implying a ~20% chance it pays $1 (plus some other possibilities). By midnight, extrapolated expected value will be $4.19, implying a ~100% chance to pay $1, plus ~45% chance that they'll make good on the suggestion of paying $10 instead, plus some other possibilities...

comment by hath · 2022-04-01T13:36:08.008Z · LW(p) · GW(p)

Wait, did people actually subscribe to the LessWrong Substack?

comment by hath · 2022-04-01T19:44:37.874Z · LW(p) · GW(p)

I've been working on setting up a TED talk at my high school, and since the beginning have been planning on asking for speakers through a post here. However, the day that we finally finished the website, and I can finally post here about it, is... when we're doing this whole GoodHeart thing. Not sure whether I should publish it today or tomorrow. (Pros: money. Cons: possibly fewer views because of everything else posted today.) What do you all think?

Replies from: lsusr
comment by lsusr · 2022-04-02T00:04:45.153Z · LW(p) · GW(p)

Pro: There's not just more posts on LW than usual. There's also way more eyeballs on LW than usual.

Con: Readers might think your TED talk is an April Fool's Day prank.

comment by hath · 2022-04-01T13:34:48.055Z · LW(p) · GW(p)

The link to "a dire spot" links to this post itself.

comment by Chris_Leong · 2022-04-03T09:30:21.754Z · LW(p) · GW(p)

I think the strongest incentive is not the financial payout, but the desire to be on the leaderboard.

comment by aphyer · 2022-04-01T18:30:17.364Z · LW(p) · GW(p)

Oh no! Our Dark Conspiracy is being foiled!

Replies from: lsusr
comment by lsusr · 2022-04-02T00:06:16.394Z · LW(p) · GW(p)

If you're confused by aphyer's comment, do a Ctrl-f for other comments on this thread that involve aphyer.

comment by Pablo Repetto (pablo-repetto-1) · 2022-04-02T19:39:12.978Z · LW(p) · GW(p)

Sorry in advance for an entirely too serious comment to a lighthearted post; it made me have thoughts I thought worth sharing. The whole "Karma convertibility" system is funny, but the irony feels slightly off. Society (vague term alert!) does in fact reward popular content with money. Goodhart's law is not "monetizing an economy instantly crashes it". My objections to Karma convertibility, are:

  1. Exploitability. Put in enough people on a system, and some of them will try to break it. Put in even more people, and they will likely succeed.
  2. Inefficiency. There is a big upfront cost to generating good ideas. Ie: you've got to invest putting interesting things inside your mind before you can make up your own. Specialization pays off just as much in intellectual labor as in other areas.
  3. Pareto's law. A small, selected group can be a consistent performance outlier.
  4. Marginal returns on money. The marginal value of money goes down dramatically when you get above the subsistence threshold.

I think that providing a salaried position to a small number of intrinsically motivated individuals is a much more efficient way of buying ideas. I think RAND is basically structured this way?

Epistemic status: Don't quote me on any of this. I've done no research, instead I'm pattern-matching from stuff that I've passively absorbed.

comment by Yitz (yitz) · 2022-04-01T20:47:41.174Z · LW(p) · GW(p)

First of all, happy April Fools!! Second, taking this seriously (🙃), I predict (contingent on this not disappearing tomorrow) that this will lead to far more posts and comments than we’ve previously seen. Will those posts and comments be higher quality? …probably not, but it will look great for SEO! Overall, I’d say that the likelihood that I’ve written this entire comment just to receive some sweet sweet good heart tokens is…~1?

comment by kave · 2022-04-01T17:43:44.982Z · LW(p) · GW(p)

(This solely applies to all new content on the site.)

Heartbreaking CDT. I’ve got a Transparent Newcomb’s I’d like to sell you

comment by Delete account (Karolina) · 2022-04-01T09:07:45.016Z · LW(p) · GW(p)

Is this an April fools joke?

Replies from: niplav, Charlie Steiner
comment by niplav · 2022-04-01T09:59:38.105Z · LW(p) · GW(p)

Do you think the Lesswrong team would lie [LW · GW] to you‽

comment by Thomas Kwa (thomas-kwa) · 2022-04-01T20:34:09.472Z · LW(p) · GW(p)

So, what are the actual biggest problems you've noticed with the LW karma system?

comment by Dagon · 2022-04-01T15:52:40.141Z · LW(p) · GW(p)

This is brilliant!  Way better than the points we used at Camp Bell long ago.  

comment by FCCC · 2022-04-08T00:47:25.055Z · LW(p) · GW(p)

I don't think this experiment could prove anything other than "it doesn't work". It's too gameable. Even if it works in the short term, that's only true for the current population. You'd change the people who join the community in the long run towards people who are willing to game the system.

comment by bfinn · 2022-04-03T10:33:41.224Z · LW(p) · GW(p)

I genuinely can’t tell how much of this is an April Fool joke. If all of it, it’s gone on too long now

Replies from: Yoav Ravid
comment by Yoav Ravid · 2022-04-03T10:43:49.222Z · LW(p) · GW(p)

It's not a joke (though it is a pun), It's a playful experiment with real money.

Replies from: habryka4
comment by habryka (habryka4) · 2022-04-04T04:51:40.373Z · LW(p) · GW(p)

Well, it's kind of a joke that it was ambiguous whether it was a joke or an experiment. Now it's more clear it's more like an experiment, but there was a substantial joke component to it.

comment by WilliamKiely · 2022-04-02T16:07:54.369Z · LW(p) · GW(p)

Did anyone with >25 Good Heart Tokens get paid out at midnight last night? I'm still unsure whether this was just an April Fool's joke or not. https://manifold.markets/WilliamKiely/will-lesswrong-pay-users-1-for-each

Replies from: aphyer
comment by aphyer · 2022-04-02T17:55:37.560Z · LW(p) · GW(p)

See here [LW · GW].

The 'midnight' thing was apparently reconsidered and removed soon after it was posted, but not before some people saw it.  Currently it looks like they will get paid out, but in around a week.

Edited to add: found the comment about the payout time, see here [LW(p) · GW(p)].

comment by AprilSR · 2022-04-02T07:19:21.604Z · LW(p) · GW(p)

I swear this post said payouts would go out at 11:59pm PST earlier, but that seems to be removed and I haven’t received my money yet. What gives?

oh new post

Replies from: Benito
comment by Ben Pace (Benito) · 2022-04-02T07:23:04.617Z · LW(p) · GW(p)

It did say that earlier, this was my mistake due to my asking an editor I know to help polish the post at 1:30am, they added the line and I didn't catch it. That line was committing me to a lot of work on a daily basis that I am not planning to do, and I removed the line when I saw it midday.

Payouts will happen in one lump sum sometime after Good Heart Week is over.

comment by Towards_Keeperhood (Simon Skade) · 2022-04-01T20:04:57.084Z · LW(p) · GW(p)

What's up with the leaderboard? Did you make a downvote worth 5$ or so, just for fun? Or what?

Replies from: gworley
comment by Gordon Seidoh Worley (gworley) · 2022-04-01T20:30:38.872Z · LW(p) · GW(p)

Based on the pattern of voting I've noticed over the past few hours, I suspect one or more people are engaged in aggressive downvoting of others. The natural people to do this would be those who want to rise to the top of the leaderboard. Since all votes are anonymous there's no good way to figure out who might be doing that.

The obvious thing to do would be to coordinate against them, but if we can't identify them then we can't do that. Seems like a possible failure of the voting system created by attaching more than internet points to them. Luckily this is all just for fun! :-)

Replies from: abramdemski
comment by abramdemski · 2022-04-01T20:35:04.002Z · LW(p) · GW(p)

I sent out a call for people to help downvote blatant exploitation, while upvoting good-faith engagement.

Replies from: Simon Skade
comment by Towards_Keeperhood (Simon Skade) · 2022-04-01T20:46:28.019Z · LW(p) · GW(p)

Well, on the leaderbord (that I see), aphyer [LW · GW] is at the top with $557, and when you click on the user and look at the votes, he almost only received downvotes. John Wentworth also received a lot of downvotes. Thus my hypothesis that a downvote is somehow worth something like $5 or so. If that is so, your call might have backfired. xD
(Though it could also be a hack or so.)

Replies from: abramdemski
comment by abramdemski · 2022-04-01T20:58:12.815Z · LW(p) · GW(p)

It looks like they were probably upvoting each other's shared drafts.

Replies from: Simon Skade
comment by hath · 2022-04-01T20:02:28.939Z · LW(p) · GW(p)

Created a market on Manifold to see if either today's GoodHeart system will last past today, or else if LW will try financial rewards for posting in 2022.

Replies from: Vaniver
comment by Vaniver · 2022-04-01T20:04:58.910Z · LW(p) · GW(p)

If you're counting the book review, are you also counting the annual review? [We don't commit to running it each year, but Laplace's Rule of Succession says the odds of doing it in 2022 are pretty high.]

Replies from: hath
comment by hath · 2022-04-01T20:14:31.865Z · LW(p) · GW(p)

...I forgot about the annual review. I think I'll just say that doesn't count, and also commit to no more changes of the conditions.

EDIT: actually, just going to kill the market.

comment by Heron (jane-mccourt) · 2022-04-01T11:27:40.859Z · LW(p) · GW(p)

Any monetisation could add to the funds of an agreed /a couple of agreed just causes. This, as opposed to individual acquisition of Good Heart Tokens, seems Goodest to me:)

Replies from: Benito
comment by Ben Pace (Benito) · 2022-04-01T17:17:34.420Z · LW(p) · GW(p)

While the form suggests using a PayPal address, you may also add an Ethereum address, or the name of a charity that you'd like us to donate it to.

comment by Unnamed · 2022-04-01T08:59:52.509Z · LW(p) · GW(p)

test post

Replies from: Chris_Leong
comment by Chris_Leong · 2022-04-01T09:11:40.028Z · LW(p) · GW(p)

Haha, I wish I thought of that idea. I know it's a joke, but still...

Replies from: Chris_Leong, Frederik
comment by Chris_Leong · 2022-04-03T09:27:56.933Z · LW(p) · GW(p)

I guess I'm glad I didn't think of this idea.

comment by Tom Lieberum (Frederik) · 2022-04-01T09:16:31.659Z · LW(p) · GW(p)

Considering you get at least one free upvote from posting/commenting itself, you just have to be faster than the downvoters to generate money :P

Replies from: Benito
comment by Ben Pace (Benito) · 2022-04-01T09:17:52.564Z · LW(p) · GW(p)

No, self-votes are not counted!

comment by Tofly · 2022-04-01T15:06:11.370Z · LW(p) · GW(p)

I will strongly upvote all replies to this comment which state that they strongly upvoted this comment.

Yes, it is easy for you to defect here, or undo your strong upvote after commenting.  However, I modeled the typical lesswrong user before making this comment, and would not have made it if they were likely to defect.

Replies from: gjm, AprilSR, gworley, johnswentworth
comment by gjm · 2022-04-01T15:40:07.401Z · LW(p) · GW(p)

I strongly upvoted the comment above.

(Then I retracted my upvote.)

Replies from: Tofly
comment by Tofly · 2022-04-01T15:44:16.579Z · LW(p) · GW(p)

I strongly upvoted the comment above.

(Then I retracted my upvote.)

Replies from: gjm
comment by gjm · 2022-04-01T18:29:33.952Z · LW(p) · GW(p)

For the avoidance of doubt, that is the response I was expecting. :-)

comment by AprilSR · 2022-04-01T15:19:06.151Z · LW(p) · GW(p)

This is a good lesson about game theory. Strong upvote.

comment by Gordon Seidoh Worley (gworley) · 2022-04-01T19:37:45.599Z · LW(p) · GW(p)

Strongly upvoted, but I recommend you not bother with voting on this comment, and instead spend your votes helping to fuel the Good Heart donation lottery [LW · GW]. Let's take LW for all its got and donate it to a good cause!

comment by aphyer · 2022-04-01T13:34:06.549Z · LW(p) · GW(p)

If anyone is up for a reciprocal trade, where I upvote all your posts in exchange for you upvoting all of mine, please let me know below!

Replies from: Measure, bfinn, johnswentworth, aphyer, johnswentworth, johnswentworth, hath, johnswentworth, johnswentworth
comment by Measure · 2022-04-01T19:41:51.471Z · LW(p) · GW(p)

Most of the child comments have 5-7 votes and near-zero net karma hearts. This suggests that non-participants are coordinating to keep the totals near zero. I'd love to participate, but I don't expect it to be worthwhile.

comment by bfinn · 2022-04-04T12:39:09.405Z · LW(p) · GW(p)

Maybe users who police bad comments/articles (like that one) should be rewarded for downvoting them by earning the $1 saved, thus maintaining financial equilibrium

Replies from: aphyer, thomas-kwa
comment by aphyer · 2022-04-04T15:05:29.919Z · LW(p) · GW(p)

Strong-downvoted, earned $6 by doing so.

(I did not actually do this)

comment by Thomas Kwa (thomas-kwa) · 2022-04-04T13:02:53.758Z · LW(p) · GW(p)

Then who decides whether the comment is actually bad?

comment by johnswentworth · 2022-04-01T18:39:35.735Z · LW(p) · GW(p)

I think we are now properly defeated. The fundamental problem is that there's only two of us mutually benefitting, so any group with more Karma than one of us can shut down the party.

Replies from: abramdemski, gworley
comment by abramdemski · 2022-04-01T18:46:43.107Z · LW(p) · GW(p)

While I would prefer for the forces of Social Coordination to win here, I would prefer they win against a real fight. Giving up so easily? Did you think for even five minutes? Shut up and do the impossible!!

comment by Gordon Seidoh Worley (gworley) · 2022-04-01T19:35:11.634Z · LW(p) · GW(p)

I'll note that the leaderboard doesn't seem to notice the downvotes and so is still showing you and aphyer in the lead. This is likely a bug but at least you're getting to have the goodest of hearts for a while for your effort. :-)

Replies from: johnswentworth
comment by johnswentworth · 2022-04-01T19:45:35.898Z · LW(p) · GW(p)

Oh, it definitely noticed the downvotes. We got about 100 karma each out of this thread, and then we lost it all and then some once people came along and started downvoting. We were both off the leaderboard entirely for a while after that.

Replies from: abramdemski, gworley
comment by abramdemski · 2022-04-01T19:59:28.610Z · LW(p) · GW(p)

But what's happening now?!? How is Aphyer suddenly at over $500 when most of their comments are now negative, and the positive ones not very high?

How?????

Replies from: johnswentworth, Taleuntum
comment by johnswentworth · 2022-04-01T20:05:27.406Z · LW(p) · GW(p)

While I would prefer for the forces of Social Coordination to win here, I would prefer they win against a real fight. Giving up so easily? Did you think for even five minutes? Shut up and do the impossible!!

comment by Taleuntum · 2022-04-01T20:36:34.249Z · LW(p) · GW(p)

My guess would be that they are upvoting some kind of post/comment which only shows up for those it is shared with. Some sites do allow these, though I'm not sure how to do this on lesswrong. I've already strong downvoted all of their public comments containing no value, so I guess they've won this time :(

EDIT: Looking around the UI I see that you can share drafts with people, maybe they're upvoting each other's drafts? (If that's even possible)

Replies from: abramdemski
comment by abramdemski · 2022-04-01T20:50:50.093Z · LW(p) · GW(p)

Don't give up yet!!!

  1. We can probably get the LW team to stop counting whatever they're pulling.
  2. We can recruit more people to try and balance things out.

EDIT: It appears something is happening... the leaderboards are shifting again...

Replies from: Taleuntum
comment by Taleuntum · 2022-04-01T20:53:11.454Z · LW(p) · GW(p)

Unfortunately, I don't have the capacity to deal with this problem right now however serious, because I have much more pressing issues in the form of r/place.

comment by Gordon Seidoh Worley (gworley) · 2022-04-01T19:54:59.612Z · LW(p) · GW(p)

Oh yeah, I started getting downvotes and now I see I've fallen quite a bit. I just noticed you two rose up quite high and I wondered what happened there!

Replies from: johnswentworth
comment by johnswentworth · 2022-04-01T19:57:07.670Z · LW(p) · GW(p)

We turned to the Light Side and gained a ton of karma by writing high-quality posts and comments, which were legitimately upvoted.

Replies from: abramdemski
comment by abramdemski · 2022-04-01T20:14:07.193Z · LW(p) · GW(p)

If so, then where are they???

I've been looking...

comment by aphyer · 2022-04-01T18:22:35.329Z · LW(p) · GW(p)

Additional additional comments

Replies from: johnswentworth
comment by johnswentworth · 2022-04-01T18:27:25.709Z · LW(p) · GW(p)

Additional additional addition, which is additive.

Replies from: aphyer
comment by aphyer · 2022-04-01T18:28:59.930Z · LW(p) · GW(p)

Multiplicative comments?

Replies from: johnswentworth
comment by johnswentworth · 2022-04-01T18:30:17.499Z · LW(p) · GW(p)

I suppose if one were to put two comments in response to each comment... but that would become rate-limited very quickly...

Replies from: aphyer, aphyer
comment by aphyer · 2022-04-01T18:31:44.293Z · LW(p) · GW(p)

Would it really?

Replies from: johnswentworth
comment by aphyer · 2022-04-01T18:32:01.521Z · LW(p) · GW(p)

Would it really?

Replies from: johnswentworth
comment by johnswentworth · 2022-04-01T18:08:47.189Z · LW(p) · GW(p)

Upvoting posts will not work unless they're new today, need to create new comments

Replies from: aphyer
comment by aphyer · 2022-04-01T18:26:19.403Z · LW(p) · GW(p)

I did not know that, thank you! I think this Legitimately Very Good Comment is one that Legitimately Deserves a strong upvote!

Replies from: johnswentworth
comment by johnswentworth · 2022-04-01T18:28:04.046Z · LW(p) · GW(p)

Indeed, an excellent point!

Replies from: aphyer
comment by aphyer · 2022-04-01T18:29:25.416Z · LW(p) · GW(p)

As does this one!

Replies from: johnswentworth
comment by johnswentworth · 2022-04-01T18:31:36.237Z · LW(p) · GW(p)

Yeah, that synergizes really well with your previous point.

Replies from: aphyer
comment by aphyer · 2022-04-01T18:33:19.232Z · LW(p) · GW(p)

Indeed! This is a high value conversation providing lots of positive sum value to all readers!

Replies from: johnswentworth
comment by johnswentworth · 2022-04-01T18:07:46.919Z · LW(p) · GW(p)

So many comments, must parallelize

Replies from: aphyer, aphyer, aphyer, aphyer, aphyer
comment by aphyer · 2022-04-01T18:25:16.570Z · LW(p) · GW(p)

Continuing to comment!

Replies from: johnswentworth
comment by johnswentworth · 2022-04-01T18:25:44.511Z · LW(p) · GW(p)

An excellent point good sir.

Replies from: aphyer
comment by aphyer · 2022-04-01T18:27:47.902Z · LW(p) · GW(p)

You likewise, my good fellow!

Replies from: johnswentworth
comment by aphyer · 2022-04-01T18:23:25.586Z · LW(p) · GW(p)

(I feel like, as the founding member of the Dark Conspiracy and the one whose parent comment is exposed to social enforcement, I should be getting more of my comments upvoted)

Replies from: johnswentworth
comment by johnswentworth · 2022-04-01T18:25:26.646Z · LW(p) · GW(p)

Eh, we'll hit the cap pretty fast regardless.

Replies from: aphyer
comment by aphyer · 2022-04-01T18:27:26.578Z · LW(p) · GW(p)

Wait, what cap?

Replies from: johnswentworth
comment by johnswentworth · 2022-04-01T18:29:08.754Z · LW(p) · GW(p)

There is currently a cap of 600 Good Heart Tokens per user while we are rolling out the system.

Replies from: aphyer
comment by aphyer · 2022-04-01T18:30:38.866Z · LW(p) · GW(p)

Oh no! Our Dark Conspiracy is being foiled!

Replies from: johnswentworth
comment by johnswentworth · 2022-04-01T18:32:23.461Z · LW(p) · GW(p)

We must find some way to escape these limits...

Replies from: aphyer
comment by aphyer · 2022-04-01T18:33:57.784Z · LW(p) · GW(p)

Perhaps by building some kind of artificial intelligence capable of faster than human clicking?

Replies from: abramdemski
comment by abramdemski · 2022-04-01T18:39:27.107Z · LW(p) · GW(p)

That would probably get you banned on perfectly normal LW moderation type grounds?

Replies from: aphyer
comment by aphyer · 2022-04-01T18:44:25.803Z · LW(p) · GW(p)

Sssssh! Don't tell them!

comment by aphyer · 2022-04-01T18:21:52.190Z · LW(p) · GW(p)

Comment all of the things!

Replies from: johnswentworth
comment by aphyer · 2022-04-01T18:22:12.669Z · LW(p) · GW(p)

Yet more comments!

Replies from: johnswentworth
comment by aphyer · 2022-04-01T18:08:39.502Z · LW(p) · GW(p)

More comments!

Replies from: aphyer, johnswentworth
comment by aphyer · 2022-04-01T18:21:28.367Z · LW(p) · GW(p)

Comments! There is a 5 second delay enforced between comments! I never knew that before!

Replies from: johnswentworth
comment by johnswentworth · 2022-04-01T18:09:10.270Z · LW(p) · GW(p)

Yes comments yay!

Replies from: aphyer
comment by aphyer · 2022-04-01T18:20:59.423Z · LW(p) · GW(p)

And more! Now that the parent is down voted no one will see them except us Conspirators! Hahahahaha!

Replies from: Vaniver, johnswentworth
comment by Vaniver · 2022-04-01T18:30:06.152Z · LW(p) · GW(p)

Recent Discussion doesn't hide children of hidden comments. I'm not sure whether this is a bug or a feature.

Replies from: aphyer
comment by aphyer · 2022-04-01T18:35:27.185Z · LW(p) · GW(p)

Oh no! They're onto us! Comment faster!!

comment by johnswentworth · 2022-04-01T18:24:29.464Z · LW(p) · GW(p)

That's a great comment!

comment by hath · 2022-04-01T13:38:04.015Z · LW(p) · GW(p)

Sure. I'll strong upvote the first 20 or so on your page.

Replies from: Lanrian, hath
comment by Lukas Finnveden (Lanrian) · 2022-04-01T15:41:37.760Z · LW(p) · GW(p)

From the post:

only votes on new content will count

Upvoting comments/posts that were made before today doesn't get you any tokens.

Replies from: hath, Viliam
comment by hath · 2022-04-01T15:47:14.060Z · LW(p) · GW(p)

Oh. Oops.

comment by Viliam · 2022-04-01T18:52:14.059Z · LW(p) · GW(p)

Just delete all old content and repost it again.

comment by hath · 2022-04-01T13:44:51.623Z · LW(p) · GW(p)

It's done. They're all strong upvoted. I keep hearing this voice, though:

“Any who disrespect the omniscient idols by misusing their knowledge for sordid financial gain will, after their death, be sent to the bottom-most layer of Hell, where venomous worms will gnaw at their organs from the inside forever, never to know rest or surcease from pain.” --Scott Alexander, Idol Words

Replies from: aphyer
comment by aphyer · 2022-04-01T13:47:21.443Z · LW(p) · GW(p)

Acknowledged, reciprocated.  If anyone else wants to join our Exploitative Karma Circle, let us know!

And I kinda hope whoever decided to do this for April Fool's Day has a plan for reverting karma changes at the end of the day, otherwise this may cause some longer-term screwups.

Replies from: hath
comment by hath · 2022-04-01T14:22:15.465Z · LW(p) · GW(p)

Yeah; my participation in this scheme was predicated on "Lightcone is occupied by smart people who 100% knew that this would be a result of today's April Fool's Day; therefore, playing along with the game is okay. Also, they can afford whatever comes as a consequence of today's thing, provided nobody actually writes a bot (something I very briefly considered before deciding it'd be too much)."

Replies from: hath
comment by hath · 2022-04-01T14:23:00.459Z · LW(p) · GW(p)

Also, I guess it's worth enabling karma change notifications for today only. Trust, but verify. Edit: looking at all the karma change as a result of this already, I really hope they plan to revert it.