Open Thread Spring 2024

post by habryka (habryka4) · 2024-03-11T19:17:23.833Z · LW · GW · 76 comments

If it’s worth saying, but not worth its own post, here's a place to put it.

If you are new to LessWrong, here's the place to introduce yourself. Personal stories, anecdotes, or just general comments on how you found us and what you hope to get from the site and community are invited. This is also the place to discuss feature requests and other ideas you have for the site, if you don't want to write a full top-level post.

If you're new to the community, you can start reading the Highlights from the Sequences, a collection of posts about the core ideas of LessWrong.

If you want to explore the community more, I recommend reading the Library [? · GW], checking recent Curated posts [? · GW], seeing if there are any meetups in your area [? · GW], and checking out the Getting Started [? · GW] section of the LessWrong FAQ [? · GW]. If you want to orient to the content on the site, you can also check out the Concepts section [? · GW].

The Open Thread tag is here [? · GW]. The Open Thread sequence is here [? · GW].

76 comments

Comments sorted by top scores.

comment by Mateusz Bagiński (mateusz-baginski) · 2024-03-15T13:22:43.444Z · LW(p) · GW(p)

Does anybody know what happened to Julia Galef?

Replies from: lahwran
comment by the gears to ascension (lahwran) · 2024-04-06T19:26:31.022Z · LW(p) · GW(p)

The only thing I can conclude looking around for her is that she's out of the public eye. Hope she's ok, but I'd guess she's doing fine and just didn't feel like being a public figure anymore. Interested if anyone can confirm that, but if it's true I want to make sure to not pry.

comment by Anand Baburajan (anand-baburajan) · 2024-03-13T18:30:22.464Z · LW(p) · GW(p)

Hello! I'm building an open source communication tool with a one-of-a-kind UI for LessWrong kind of deep, rational discussions. The tool is called CQ2 (https://cq2.co). It has a sliding panes design with quote-level threads. There's a concept of "posts" for more serious discussions with many people and there's "chat" for less serious ones, and both of them have a UI crafted for deep discussions.

I simulated some LessWrong discussions there – they turned out to be a lot more organised and easy to follow. You can check them out in the chat channel and direct message part of the demo on the site. However, it is a bit inconvenient – there's horizontal scrolling and one needs to click to open new threads. Since forums need to prioritize convenience, I think CQ2's design isn't good for LessWrong. But I think the inconvenience is worth it for such discussions at writing-first teams, since it helps them with hyper-focusing on one thing at a time and avoid losing context in order to come to a conclusion and make decisions.

If you have such discussions at work, I would love to learn about your team, your frustrations with existing communication tools, and better understand how CQ2 can help! I would appreciate any feedback or leads! I feel my comment might come off as an ad, but I (and CQ2) strongly share LessWrong's "improving our reasoning and decision-making" core belief and it's open source.

I found LessWrong a few months back. It's a wonderful platform and I particularly love the clean design. I've always loved how writing forces a deeper clarity of thinking and focuses on getting to the right answer.

P.S. I had mistakenly posted this comment in the previous, old Open Thread, hence resharing here.

Replies from: papetoast, anand-baburajan, habryka4, Celarix
comment by papetoast · 2024-04-21T08:08:41.697Z · LW(p) · GW(p)

I just stumbled on this website: https://notes.andymatuschak.org/About_these_notes It has a similar UI but for Obsidian-like linked notes. The UI seem pretty good.

Replies from: anand-baburajan
comment by Anand Baburajan (anand-baburajan) · 2024-04-23T09:44:27.857Z · LW(p) · GW(p)

I like his UI. In fact, I shared about CQ2 with Andy in February since his notes site was the only other place where I had seen the sliding pane design. He said CQ2 is neat!

comment by Anand Baburajan (anand-baburajan) · 2024-03-27T05:33:18.365Z · LW(p) · GW(p)

Update: now you can create discussions on CQ2! And, here's a demo with an actual LessWrong discussion between Vanessa and Rob: https://cq2.co/demo.

comment by habryka (habryka4) · 2024-03-13T20:26:47.048Z · LW(p) · GW(p)

This is cool!

Two pieces of feedback: 

  1. I think it's quite important that I can at least see the number of responses to a comment before I have to click on the comment icon. Currently it only shows me a generic comment icon if there are any replies.
  2. I think one of the core use-cases of a comment UI is reading back and forth between two users. This UI currently makes that a quite disjointed operation. I think it's fine to prioritize a different UI experience, but it does feel like a big loss to me.
Replies from: anand-baburajan
comment by Anand Baburajan (anand-baburajan) · 2024-03-14T15:30:36.084Z · LW(p) · GW(p)

Thanks for the feedback!

I think it's quite important that I can at least see the number of responses to a comment before I have to click on the comment icon. Currently it only shows me a generic comment icon if there are any replies.

Can you share why you think it's quite important (for a work communication tool)? For a forum, I think it would make sense -- many people prefer reading the most active threads. For a work communication tool, I can't think of any reason why it would matter how many comments a thread has.

I think one of the core use-cases of a comment UI is reading back and forth between two users. This UI currently makes that a quite disjointed operation. I think it's fine to prioritize a different UI experience, but it does feel like a big loss to me.

I thought about this for quite a while and have started to realise that the "posts" UI could be too complicated. I'm going to try out the "chat" and "DMs" UI for posts and see how it goes. Thanks!

Although "Chat" and "DMs"' UI allows easily followable back and forth between people, I would like to point out that CQ2 advocates for topic-wise discussions, not person-wise. Here [LW(p) · GW(p)]'s an example comment from LessWrong. In that comment, it's almost impossible to figure out where the quotes are from -- i.e., what's the context. And what happened next is another person replied to that comment with more quotes. This example was a bit extreme with many quotes but I think my point applies to every comment with quotes. One needs to scroll person-wise through so many topics, instead of topic-wise. I (and CQ2) prefer exploring what are people's thoughts topic-by-topic, not what are the thoughts on all topics simultaneously, person-by-person.

Again, not saying my design is good for LessWrong; I understand forums have their own place. But I think for a tool for work, people would prefer topic-wise over person-wise.

Replies from: jmh, anand-baburajan
comment by jmh · 2024-03-15T16:46:23.935Z · LW(p) · GW(p)

My sense, regarding the read the most active thread desire, is that the most active thread might well be amongst either the team working on some project under discussion or across teams that are envolved in or impacted by some project. In such a case I would think knowing where the real discussion is taking place regarding some "corporate discussions" might be helpful and wanted.

I suppose the big question there is what about all the other high volume exchanges, are they more personality driven rather than subject/substance driven. Does the comment count just be a really noisy signal to try keying off?

comment by Anand Baburajan (anand-baburajan) · 2024-03-14T17:32:08.594Z · LW(p) · GW(p)

P.S. I'm open to ideas on building this in collaboration with LessWrong!

comment by Celarix · 2024-04-03T14:26:03.566Z · LW(p) · GW(p)

Ooh, nice. I've been wanting this kind of discussion software for awhile. I do have a suggestion: maybe, when hovering over a highlighted passage, you could get some kind of indicator of how many child comments are under that section, and/or change the highlight contrast for threads that have more children, so we can tell which branches of the discussion got the most attention

Replies from: anand-baburajan
comment by Anand Baburajan (anand-baburajan) · 2024-04-05T05:53:28.858Z · LW(p) · GW(p)

Thanks @Celarix [LW · GW]! I've got the same feedback from three people now, so seems like a good idea. However, I haven't understood why it's necessary. For a forum, I think it would make sense -- many people prefer reading the most active threads. For a discussion tool, I can't think of any reason why it would matter how many comments a thread has. Maybe the point is to let a user know if there's any progress in a thread over time, which makes sense.

Replies from: Celarix
comment by Celarix · 2024-04-05T15:14:02.010Z · LW(p) · GW(p)

My thinking is that the more discussed threads would have more value to the user. Small threads with 1 or 2 replies are more likely to be people pointing out typos or just saying +1 to a particular passage.

Of course, there is a spectrum - deeply discussed threads are more likely to be angry back-and-forths that aren't very valuable.

Replies from: anand-baburajan
comment by Anand Baburajan (anand-baburajan) · 2024-04-08T06:22:01.247Z · LW(p) · GW(p)

would have more value to the user

This feels self and learning focused, as opposed to problem and helping focused, and I'm building CQ2 for the latter.

Small threads with 1 or 2 replies are more likely to be people pointing out typos or just saying +1 to a particular passage.

There could also be important and/or interesting points in a thread with only 1 or 2 replies, and implementing this idea would prevent many people from finding that point, right?

just saying +1 to a particular passage

Will add upvote/downvote.

comment by atergitna (greta-goodwin) · 2024-04-03T19:29:17.991Z · LW(p) · GW(p)

Hi! I have been lurking here for over a year but I've been too shy to participate until now. I'm 14, and I've been homeschooled all my life. I like math and physics and psychology, and I've learned lots of interesting things here. I really enjoyed reading the sequences last year. I've also been to some meetups in my city and the people there (despite – or maybe because of – being twice my age) are very cool. Thank you all for existing!

Replies from: kave
comment by kave · 2024-04-03T20:23:25.069Z · LW(p) · GW(p)

Hello and welcome to the site! I'm glad you're saying hello despite having been too shy :-)

Do let us know in this thread or in the intercom in the bottom right if you run into any problems.

comment by thornoar · 2024-04-19T17:43:57.819Z · LW(p) · GW(p)

Hello everyone! My name is Roman Maksimovich, I am an immigrant from Russia, currently finishing high school in Serbia. My primary specialization is mathematics, and back in middle school I have had enough education in abstract mathematics (from calculus to category theory and topology) to call myself a mathematician.

My other strong interests include computer science and programming (specifically functional programming, theoretical CS, AI, and systems programming s.a. Linux) as well as languages (specifically Asian languages like Japanese).

I ended up here after reading HP:MOR, which I consider to be an all-time masterpiece. The Sequences are very good too, although not that gripping. Rationality is a very important principle in my life, and so far I found the forum to be very well-organized and the posts to be very informative and well-written, so I will definitely stick around and try to engage in the forum to the best of my ability.

I thought I might do a bit of self-advertising as well. Here's my GitHub: https://github.com/thornoar

If any of you use this very niche mathematical graphics tool called Asymptote, you might be interested to know that I have been developing a cool 6000-line Asymptote library called 'smoothmanifold', which is sort of like a JavaScript framework (an analogy that I do not like) but for drawing abstract mathematical diagrams with Asymptote, whose main problem is the lack of abstraction. In plain Asymptote, you usually have to specify all the coordinates manually and draw objects line by line. In my library I make it so that the code resembles the logical structure of the picture more. You can draw a set as a blob on the plane, and then draw arrows that connect different sets, which would be a nightmare to do manually. And this is only the beginning -- there is a lot more features. If any of this is what interests you, feel free to read the README.md.

I have also written some mathematical papers, the most recent one paired with a software program for strong password creation. If you are interested in cryptography and cybersecurity, and would to create strong passwords using a hashing algorithm, you can take a look at 'password-hash', which contains both the algorithm source and binaries, as well as the paper/documentation.

comment by complicated.world (w_complicated) · 2024-03-28T01:29:59.362Z · LW(p) · GW(p)

Hi LessWrong Community!

I'm new here, though I've been an LW reader for a while. I'm representing complicated.world website, where we strive to use similar rationality approach as here and we also explore philosophical problems. The difference is that, instead of being a community-driven portal like you, we are a small team which is working internally to achieve consensus and only then we publish our articles. This means that we are not nearly as pluralistic, diverse or democratic as you are, but on the other hand we try to present a single coherent view on all discussed problems, each rooted in basic axioms. I really value the LW community (our entire team does) and would like to start contributing here. I would also like to present from time to time a linkpost from our website - I hope this is ok. We are also a not-for-profit website.

Replies from: habryka4
comment by habryka (habryka4) · 2024-03-28T04:17:15.606Z · LW(p) · GW(p)

Hey! 

It seems like an interesting philosophy. Feel free to crosspost. You've definitely chosen some ambitious topics to try to cover, which I am generally a fan of.

Replies from: w_complicated
comment by complicated.world (w_complicated) · 2024-03-28T13:03:53.016Z · LW(p) · GW(p)

Thanks! The key to topic selection is where we find that we are most disagreeing with the popular opinions. For example, the number of times I can cope with hearing someone saying "I don't care about privacy, I have nothing to hide" is limited. We're trying to have this article out before that limit is reached. But in order to reason about privacy's utility and to ground it in root axioms, we first have to dive into why we need freedom. That, in turn requires thinking about mechanisms of a happy society. And that depends on our understanding of happiness, hence that's where we're starting.

comment by metalcrow · 2024-03-26T20:33:05.610Z · LW(p) · GW(p)

Hello! I'm dipping my toes into this forum, coming primarily from the Scott Alexander side of rationalism. Wanted to introduce myself, and share that i'm working on a post about ethics/ethical frameworks i hope to share here eventually!

Replies from: habryka4
comment by habryka (habryka4) · 2024-03-27T22:29:34.597Z · LW(p) · GW(p)

Hey metalcrow! Great to have you here! Hope you have a good time and looking forward to seeing your post!

comment by Daniel Kokotajlo (daniel-kokotajlo) · 2024-04-05T15:12:28.612Z · LW(p) · GW(p)

Feature request: I'd like to be able to play the LW playlist (and future playlists!) from LW. I found it a better UI than Spotify and Youtube, partly because it didn't stop me from browsing around LW and partly because it had the lyrics on the bottom of the screen. So... maybe there could be a toggle in the settings to re-enable it?

Replies from: habryka4
comment by habryka (habryka4) · 2024-04-05T18:50:22.897Z · LW(p) · GW(p)

I was unsure whether people would prefer that, and decided yesterday to instead cut it, but IDK, I do like it. I might clean up the code and find some way to re-activate it on the site.

Replies from: Dagon, whestler
comment by Dagon · 2024-04-06T15:43:56.675Z · LW(p) · GW(p)

I liked it, but probably don't want it there all the time.  I wonder if it's feasible (WRT your priority list) to repeat some of the site feature options from account settings on a "quick feature menu", to make it easy to turn on and off.

comment by whestler · 2024-04-10T09:52:50.230Z · LW(p) · GW(p)

In terms of my usage of the site, I think you made the right call. I liked the feature when listening but I wanted to get rid of it afterwards and found it frustrating that it was stuck there. Perhaps something hidden on a settings page would be appropriate, but I don't think it's needed as a default part of the site right now.

comment by niplav · 2024-04-22T18:51:12.456Z · LW(p) · GW(p)

There are several sequences which are visible on the profiles of their authors, but haven't yet been added to the library [? · GW]. Those are:

I think these are good enough to be moved into the library.

Replies from: habryka4
comment by habryka (habryka4) · 2024-04-22T21:08:26.706Z · LW(p) · GW(p)

This probably should be made more transparent, but the reason why these aren't in the library is because they don't have images for the sequence-item. We display all sequences that people create that have proper images on the library (otherwise we just show it on user's profiles).

comment by niplav · 2024-04-08T17:44:56.895Z · LW(p) · GW(p)

Obscure request:

Short story by Yudkowsky, on a reddit short fiction subreddit, about a time traveler coming back to the 19th century from the 21st. The time traveler is incredibly distraught about the red tape in the future, screaming about molasses and how it's illegal to sell food on the street.

Nevermind, found it.

comment by skybluecat · 2024-04-06T14:00:34.425Z · LW(p) · GW(p)

Hi! I have lurked for quite a while and wonder if I can/should participate more. I'm interested in science in general, speculative fiction and simulation/sandbox games among other stuff. I like reading speculations about the impact of AI and other technologies, but find many of the alignment-related discussions too focused on what the author wants/values rather than what future technologies can really cause. Also, any game recommendations with a hard science/AI/transhumanist theme that are truly simulation-like and not narratively railroading?

comment by jmh · 2024-03-17T13:55:25.150Z · LW(p) · GW(p)

How efficient are equity markets? No, not in the EMH sense. 

My take is that market efficiency viewed from economics/finance is about total surplus maximization -- the area between the supply and demand curves. Clearly when S and D are order schedules and P and Q correspond to the S&D intersection one maximizes the area of the triangle defined in the graph.

But existing equity markets don't work off an ordered schedule but largely match trades in a somewhat random order -- people place orders (bids and offers) throughout the day and as they come in during market hours trades occur.

Given these are pure pecuniary markets the total surplus represents something of a total profit in the market for the day's activities (clearly something different than the total profits to the actual share owners who sold so calling it profit might be a bit confusing). One might think markets should be structured to maximize that area but clearly that is not the case. 

It would be a very unsual case for the daily order flow to perfectly match with the implied day demand and supply curves that would represent the bids and offers (lets call them the "real" bids and offers but I'm not entirely sure how to distinquish that from other bids and offers that will start evaporating as soon as they become the market bid or offer). So would a settlement structure line mutual funds produce a better outcome for equities? 

Maybe. In other words, rather than putting a market order in and having it executed right then or putting a limit order in and if the market moves to that price it executes, all orders get put in the order book and then at market close the clearing price is determed and those trades that actually make sense occur. What is prevented is the case of either buyers above the clearlin price from getting paired with sellers that are also above the clearling price, or the reverse, buyers biding below the clearing price pairing with sellers who are also willing to sell below the clearin price. Elimination of both inframarginal and extramarginal trades that represend low value exchange pairings.

One thing I wonder about here is what information might be lost/masked and if the informational value might outweigh the reduction in suplus captured. But I'm also not sure that whatever information might be seen in the current structure is not also present in the end of day S&D schedules and so fully reflected in the price outcomes.

Replies from: thomas-kwa
comment by Thomas Kwa (thomas-kwa) · 2024-03-17T18:58:48.629Z · LW(p) · GW(p)

In practice it is not as bad as uniform volume throughout the day would be for two reasons:

  • Market-makers narrow spreads to prevent any low-value-exchange pairings that would be predictable price fluctuations. They do extract some profits in the process.
  • Volume is much higher near the open and close.

I would guess that any improvements of this scheme would manifest as tighter effective spreads, and a reduction in profits of HFT firms (which seem to provide less value to society than other financial firms).

Replies from: jmh
comment by jmh · 2024-03-17T19:43:01.299Z · LW(p) · GW(p)

I had prehaps a bit unjustly tossed the market maker role into that "not real bid/off" bucket. I also agree they do serve to limit the worst case matches. But such a role would simply be unnecessary so I still wonder about the cost in terms of the profits captured by the market makers. Is that a necessary cost in today's world? Not sure.

And I do say that as someone who is fairly active in the markets and have taken advantage of thin markets in the off market hours sessions where speads can widen up a lot.

comment by gilch · 2024-04-17T21:33:14.218Z · LW(p) · GW(p)

PSA: Tooth decay might be reversible! The recent discussion around the Lumina [LW · GW] anti-cavity prophylaxis reminded me of a certain dentist's YouTube channel I'd stumbled upon recently, claiming that tooth decay can be arrested and reversed using widely available over-the-counter dental care products. I remember my dentist from years back telling me that if regular brushing and flossing doesn't work, and the decay is progressing, then the only treatment option is a filling. I wish I'd known about alternatives back then, because I definitely would have tried that first. Remineralization wouldn't have helped in the case when I broke a tooth, but I maybe could have avoided all my other fillings. I am very suspicious of random health claims on the Internet, but this one seemed reasonably safe and cheap to try, even if it ultimately doesn't work.

Replies from: nathan-helm-burger, abandon
comment by Nathan Helm-Burger (nathan-helm-burger) · 2024-04-24T17:08:27.824Z · LW(p) · GW(p)

I've been using a remineralization toothpaste imported from Japan for several years now, ever since I mentioned reading about remineralization to a dentist from Japan. She recommended yhe brand to me. FDA is apparently bogging down release in the US, but it's available on Amazon anyway. It seems to have slowed, but not stopped, the formation of cavities. It does seem to result in faster plaque build-up around my gumline, like the bacterial colonies are accumulating some of the minerals not absorbed by the teeth. The brand I use is apagard.

comment by dirk (abandon) · 2024-04-18T08:39:10.440Z · LW(p) · GW(p)

I apologize for my lack of time to find the sources for this belief, so I could well be wrong, but my recollection of looking up a similar idea is that I found it to be reversible only in the very earliest stages, when the tooth has weakened but not yet developed a cavity proper.

Replies from: gilch
comment by gilch · 2024-04-18T16:40:04.135Z · LW(p) · GW(p)

I didn't say "cavity"; I said, "tooth decay". No-one is saying remineralization can repair a chipped, cracked, or caved-in tooth. But this dentist does claim that the decay (caries) can be reversed even after it has penetrated the enamel and reached the dentin, although it takes longer (a year instead of months), by treating the underlying bacterial infection and promoting mineralization. It's not clear to me if the claim is that a small hole can fill in on its own, but a larger one probably won't although the necessary dental treatment (filling) in that case will be less invasive if the surrounding decay has been arrested.

I am not claiming to have tested this myself. This is hearsay. But the protocol is cheap to try and the mechanism of action seems scientifically plausible given my background knowledge.

comment by Steven Byrnes (steve2152) · 2024-04-02T17:00:40.285Z · LW(p) · GW(p)

I’m in the market for a new productivity coach / accountability buddy, to chat with periodically (I’ve been doing one ≈20-minute meeting every 2 weeks) about work habits, and set goals, and so on. I’m open to either paying fair market rate, or to a reciprocal arrangement where we trade advice and promises etc. I slightly prefer someone not directly involved in AGI safety/alignment—since that’s my field and I don’t want us to get nerd-sniped into object-level discussions—but whatever, that’s not a hard requirement. You can reply here, or DM or email me. :) update: I’m all set now

comment by Blacknsilver · 2024-03-19T08:29:14.567Z · LW(p) · GW(p)

Post upvotes are at the bottom but user comment upvotes are at the top of each comment. Sometimes I'll read a very long comment and then have to scroll aaaaall the way back up to upvote it. Is there some reason for this that I'm missing or is it just an oversight?

Replies from: habryka4
comment by habryka (habryka4) · 2024-03-19T16:12:15.161Z · LW(p) · GW(p)

Post upvotes are both at the bottom and top, but repeating them for comments at the bottom looks a lot too cluttered. Having them at the top is IMO more important since you want to be able to tell how good something is before you read it.

comment by michael_mjd · 2024-04-25T19:17:41.756Z · LW(p) · GW(p)

Is there a post in the Sequences about when it is justifiable to not pursue going down a rabbit hole? It's a fairly general question, but the specific context is a tale as old as time. My brother, who has been an atheist for decades, moved to Utah. After 10 years, he now asserts that he was wrong and his "rigorous pursuit" of verifying with logic and his own eyes, leads him to believe the Bible is literally true. I worry about his mental health so I don't want to debate him, but felt like I should give some kind of justification for why I'm not personally embarking on a bible study. There's a potential subtext of, by not following his path, I am either not that rational, or lack integrity. The subtext may not really be there, but I figure if I can provide a well thought out response or summarize something from EY, it might make things feel more friendly, e.g. "I personally don't have enough evidence to justify spending the time on this, but I will keep an open mind if any new evidence comes up."

comment by cubefox · 2024-04-03T13:36:34.308Z · LW(p) · GW(p)

Is it really desirable to have the new "review bot" in all the 100+ karma comment sections? To me it feels like unnecessary clutter, similar to injecting ads.

Replies from: habryka4
comment by habryka (habryka4) · 2024-04-03T16:28:07.162Z · LW(p) · GW(p)

Where else would it go? We need a minimum level of saliency to get accurate markets, and I care about the signal from the markets a good amount.

Replies from: Dagon
comment by Dagon · 2024-04-06T16:05:17.241Z · LW(p) · GW(p)

I haven't noticed it (literally at all - I don't think I've seen it, though I'm perhaps wrong).  Based on this comment, I just looked at https://www.lesswrong.com/users/review-bot?from=search_autocomplete [LW · GW] and it seems a good idea (and it points me to posts I may have missed - I tend to not look at the homepage, just focusing on recent posts and new comments on posts on https://www.lesswrong.com/allPosts). [? · GW]

I think putting a comment there is a good mechanism to track, and probably easier and less intrusive than a built-in site feature.  I have no clue if you're actually getting enough participation in the markets to be useful - it doesn't look like it at first glance, but perhaps I'm wrong. 

It does seem a little weird (and cool, but mostly in the "experiment that may fail, or may work so well we use it elsewhere" way) to have yet another voting mechanism for posts.  I kind of like the explicitness of "make a prediction about the future value of this post" compared to "loosely-defined up or down".  

comment by Mateusz Bagiński (mateusz-baginski) · 2024-03-29T07:40:39.722Z · LW(p) · GW(p)

I have the mild impression that Jacqueline Carey's Kushiel trilogy is somewhat popular in the community?[1] Is it true and if so, why?

  1. ^

    E.g. Scott Alexander references Elua in Mediations on Moloch and I know of at least one prominent LWer who was a big enough fan of it to reference Elua in their discord handle.

Replies from: habryka4
comment by habryka (habryka4) · 2024-03-29T19:32:07.492Z · LW(p) · GW(p)

My model is that it's mostly popular because Scott Alexander referenced it in Meditations on Moloch, and the rest is just kind of background popularity, but not sure.

comment by HiddenPrior (SkinnyTy) · 2024-03-28T18:38:17.635Z · LW(p) · GW(p)

Unsure if there is normally a thread for putting only semi-interesting news articles, but here is a recently posted news article by Wired that seems.... rather inflammatory toward Effective Altruism. I have not read the article myself yet, but a quick skim confirms the title is not only to get clickbait anger clicks, the rest of the article also seems extremely critical of EA, transhumanism, and Rationality. 

I am going to post it here, though I am not entirely sure if getting this article more clicks is a good thing, so if you have no interest in reading it maybe don't click it so we don't further encourage inflammatory clickbait tactics. 

https://www.wired.com/story/deaths-of-effective-altruism/?utm_source=pocket-newtab-en-us

Replies from: SkinnyTy
comment by HiddenPrior (SkinnyTy) · 2024-03-28T19:53:08.437Z · LW(p) · GW(p)

I did a non-in-depth reading of the article during my lunch break, and found it to be of lower quality than I would have predicted. 

I am open to an alternative interpretation of the article, but most of it seems very critical of the Effective Altruism movement on the basis of "calculating expected values for the impact on peoples lives is a bad method to gauge the effectiveness of aid, or how you are impacting peoples lives." 

The article begins by establishing that many medicines have side effects. Since some of these side effects are undesirable, the author suggests, though they do not state explicitly, that the medicine may also be undesirable if the side effect is bad enough. They go on to suggest that Givewell, and other EA efforts at aid are not very aware of the side effects of their efforts, and that the efforts may therefore do more harm than good. The author does not stoop so low as to actually provide evidence of this, or even make any explicit claims that could be checked or contradicted, but merely suggests that givewell does not do a good job of this.

This is the less charitable part of my interpretation (no pun intended), but I feel the author spends a lot of the article constantly suggesting that trying to be altruistic, especially in an organized or systematic way, is ineffective, maybe harmful and generally not worth the effort. Mostly the author does this by suggesting anecdotal stories of their investigations into charity, and how they feel much wiser now.

The author then moves on to their association of SBF with Effective Altruism, going so far as to say: "Sam Bankman-Fried is the perfect prophet of EA, the epitome of its moral bankruptcy." In general, the author goes on to give a case for how SBF is the classic utilitarian villain, justifying his immoral acts through oh-so esoteric calculations of improving good around the world on net. 

The author goes on to lay out a general criticism of Effective Altruism as relying on arbitrary utilitarian measures of moral value, such as what counts as a life saved. The author suggests Effective Altruism has become popular because Billionaires like how it makes a straightforward case for converting wealth into moral good, and generally attempts to undermine this premise. 

The author is generally extremely critical of EA, and any effort at organized charity, and suggests that the best alternative to EA (or utilitarian moral reasoning in general, I presume) is the following:

 

the “dearest test.” When you have some big call to make, sit down with a person very dear to you—a parent, partner, child, or friend—and look them in the eyes. Say that you’re making a decision that will affect the lives of many people, to the point that some strangers might be hurt. Say that you believe that the lives of these strangers are just as valuable as anyone else’s. Then tell your dearest, “I believe in my decisions, enough that I’d still make them even if one of the people who could be hurt was you.”

Or you can do the “mirror test.” Look into the mirror and describe what you’re doing that will affect the lives of other people. See whether you can tell yourself, with conviction, that you’re willing to be one of the people who is hurt or dies because of what you’re now deciding. Be accountable, at least, to yourself.

Which I suppose is fine, but I think this reveals the author is primarily concerned about their personal role or responsibility in causing positive or negative moral events, and that the author has very little regard for a consequentialist view of the actual state of reality. Unfortunately, the author does very little do directly engage in dialogue about moral values, and makes the assumption throughout the entire article that everyone does, or at least should, share their own moral values. 

The author finishes the article with an anecdote of their friend, who they suggest is a better example of being an altruist since they fly out to an island themselves, where they provide direct aid with water stations, and the direct accountability and lack of billionaires demonstrates how selfless and good he is. 
 

I don't know who this author is, but I get the feeling they are very proud of this article, and they should surely congratulate themselves on spending their time, and the time of their readers so well. 

TL;DR
All in all, I think this article can best be summarized by honestly expressing that I feel I wasted my time reading it, and writing this summary. I considered deleting my post on this article, so that I would not risk others also wasting their time on it, but I will leave this summary up so that they can at least waste less time on this article. 

comment by jmh · 2024-03-12T20:00:07.188Z · LW(p) · GW(p)

I don't think this crosses the line regarding poltics on the board but note that as a warning header.

I was just struck by a though related to the upcoming elections in the USA. Age of both party's candidate have been noted and create both some concern and even risks for the country.

No age limits exist and I suspect trying to get get legislative action on that would be slow to impossible as it undoubtedly would be a new ammendment to the Constitution.

BUT, I don't think there is any law or other restriction on any political party imposing their own age limit for any candidate they will nomimate to Federal or State positions. If not, and if anyone knows please speak up:

  1. Would the existing incentive strucutures suggest it might be easier for a political party to do this than expecting Congress to address these concerns?
  2. Should any party enact such a rule, would that be a completitive advantage in the way of improved (at the margins at least) for that party in the competition for membership and voters?
  3. Would that in any way help to improve the performance of parties by mitigating both entrenched status quo leadership and perhaps reduce overall factionalism within the party?
Replies from: MondSemmel, james-camacho
comment by MondSemmel · 2024-03-13T15:00:04.965Z · LW(p) · GW(p)

From what I understand, because the US Electoral College is structured such that state laws determine who the electors will vote for as president, you wouldn't need any constitutional amendment or federal legislative action to impose an age limit for the US presidential election in particular. In contrast, I think the lower age limit of 35 for US presidents is a constitutional requirement, and as such would not be nearly as easy to change.

On a somewhat related note, there's an interesting attempt by US states to assign electoral votes based on the national popular vote.

In 48 of the 50 states, state laws mandate that the winner of the plurality of the statewide popular vote receive all of that state's electoral votes.

Based on this Wikipedia quote, I imagine states could impose arbitrary requirements for who can or cannot receive the electoral votes, including imposing an age limit. Basically, add a clause to the state laws that "Electors must abstain if the winner of the plurality does not fulfill the following requirements...".

EDIT: Note, however, that if no candidate gets a majority of the electoral vote (270+ votes), then the US House of Representatives elects the US President instead. So while such a state law would disincentivize particular candidates, if such a candidate ran for president anyway and won the plurality of the state vote, then the abstention of the electors might well result in the Electoral College disempowering itself. And furthermore the House of Representatives could still elect an arbitrary candidate.

EDIT2: Okay, I think I've come up with a better state law design: If the winner of the plurality of state votes exceeds the age limit, then assign the electoral votes to either the second place instead (regardless of their age), or alternatively to whoever of the top two candidates is younger. Either version ensures that the electoral college will not abstain, which makes the House of Representatives route less likely. And either version disincentivizes a scenario where the presidential candidates of both parties exceed the age limit, since in this case, both parties are incentivized to run either a candidate below the age limit, or if not that, then at least a younger candidate than the opposing party's. And only the former strategy results in a stable outcome, whereas the strategy of running a younger candidate above the age limit can be circumvented by the opposing party running a still younger candidate.

comment by James Camacho (james-camacho) · 2024-04-02T22:58:02.521Z · LW(p) · GW(p)

Age limits do exist: you have to be at least 35 to run for President, at least 30 for Senator, and 25 for Representative. This automatically adds a decade or two to your candidates.

comment by Drake Morrison (Leviad) · 2024-04-17T20:18:35.953Z · LW(p) · GW(p)

Feature Suggestion: add a number to the hidden author names.

I enjoy keeping the author names hidden when reading the site, but find it difficult to follow comment threads when there isn't a persistent id for each poster. I think a number would suffice while keeping the hiddenness.

comment by Thomas Kwa (thomas-kwa) · 2024-04-12T06:50:23.453Z · LW(p) · GW(p)

How much power is required to run the most efficient superhuman chess engines? There's this discussion saying Stockfish running on a phone is superhuman, but is that one watt or 10 watts? Could we beat grandmasters with 0.1 watts if we tried?

Replies from: quetzal_rainbow
comment by quetzal_rainbow · 2024-04-12T07:49:09.101Z · LW(p) · GW(p)

It's kinda ill-formed question, because you can get the same performance if you compute moves longer with lower power. I guess you are searching for something like"energy per move".

Replies from: thomas-kwa
comment by Thomas Kwa (thomas-kwa) · 2024-04-12T08:11:10.802Z · LW(p) · GW(p)

The question makes sense if you fix a time control.

comment by Mateusz Bagiński (mateusz-baginski) · 2024-04-10T05:52:44.125Z · LW(p) · GW(p)

Any thoughts on Symbolica? (or "categorical deep learning" more broadly?)

All current state of the art large language models such as ChatGPT, Claude, and Gemini, are based on the same core architecture. As a result, they all suffer from the same limitations.

Extant models are expensive to train, complex to deploy, difficult to validate, and infamously prone to hallucination. Symbolica is redesigning how machines learn from the ground up. 

We use the powerfully expressive language of category theory to develop models capable of learning algebraic structure. This enables our models to have a robust and structured model of the world; one that is explainable and verifiable.

It’s time for machines, like humans, to think symbolically.

  1. How likely is it that Symbolica [or sth similar] produces a commercially viable product?
  2. How likely is it that Symbolica creates a viable alternative for the current/classical DL?
    1. I don't think it's that different from the intentions behind Conjecture's CoEms proposal. [LW · GW] (And it looks like Symbolica have more theory and experimental results backing up their ideas.)
      1. Symbolica don't use the framing of AI [safety/alignment/X-risk], but many people behind the project are associated with the Topos Institute that hosted some talks from e.g. Scott Garrabrant or Andrew Critch.
  3. What is the expected value of their research for safety/verifiability/etc?
    1. Sounds relevant to @davidad [LW · GW]'s plan [LW · GW], so I'd be especially curious to know his take.
  4. How likely is it that whatever Symbolica produces meaningfully contributes to doom (e.g. by advancing capabilities research without at the same time sufficiently/differentially advancing interpretability/verifiability of AI systems)?

(There's also PlantingSpace but their shtick seems to be more "use probabilistic programming and category theory to build a cool Narrow AI-ish product" whereas Symbolica want to use category theory to revolutionize deep learning.)

Replies from: D0TheMath, faul_sname
comment by Garrett Baker (D0TheMath) · 2024-04-16T23:44:42.750Z · LW(p) · GW(p)

A new update

Hi John,

thank you for sharing the job postings. We’re starting something really exciting, and as research leads on the team, we - Paul Lessard and Bruno Gavranović - thought we’d provide clarifications.

Symbolica was not started to improve ML using category theory. Instead, Symbolica was founded ~2 years ago, with its 2M seed funding round aimed at tackling the problem of symbolic reasoning, but at the time, its path to getting there wasn’t via categorical deep learning (CDL). The original plan was to use hypergraph rewriting as means of doing learning more efficiently. That approach however was eventually shown unviable.

Symbolica’s pivot to CDL started about five months ago. Bruno had just finished his Ph.D. thesis laying the foundations for the topic and we reoriented much of the organization towards this research direction. In particular, we began: a) refining a roadmap to develop and apply CDL, and b) writing a position paper, in collaboration with with researchers at Google DeepMind which you’ve cited below.

Over these last few months, it has become clear that our hunches about applicability are actually exciting and viable research directions. We’ve made fantastic progress, even doing some of the research we planned to advocate for in the aforementioned position paper. Really, we discovered just how much Taking Categories Seriously gives you in the field of Deep Learning.

Many advances in DL are about creating models which identify robust and general patterns in data (see the Transformers/Attention mechanism, for instance). In many ways this is exactly what CT is about: it is an indispensable tool for many scientists, including ourselves, to understand the world around us: to find robust patterns in data, but also to communicate, verify, and explain our reasoning.

At the same time, the research engineering team of Symbolica has made significant, independent, and concrete progress implementing a particular deep learning model that operates on text data, but not in an autoregressive manner as most GPT-style models do.

These developments were key signals to Vinod and other investors, leading to the closing of the 31M funding round.

We are now developing a research programme merging the two, leveraging insights from theories of structure, e.g. categorical algebra, as means of formalising the process by which we find structure in data. This has twofold consequence: pushing models to identify more robust patterns in data, but also interpretable and verifiable ones.

In summary:

a) The push to apply category theory was not based on a singular whim, as the the post might suggest,

but that instead

b) Symbolica is developing a serious research programme devoted to applying category theory to deep learning, not merely hiring category theorists

All of this is to add extra context for evaluating the company, its team, and our direction, which does not come across in the recently published tech articles.

We strongly encourage interested parties to look at all of the job ads, which we’ve tailored to particular roles. Roughly, in the CDL team, we’re looking for either

1) expertise in category theory, and a strong interest in deep learning, or

2) expertise in deep learning, and a strong interest in category theory.

at all levels of seniority.

Happy to answer any other questions/thoughts.

Bruno Gavranović,

Paul Lessard

comment by faul_sname · 2024-04-13T00:55:32.733Z · LW(p) · GW(p)

I'd bet against anything particularly commercially successful. Manifold could give better and more precise predictions if you operationalize "commercially viable".

Replies from: D0TheMath
comment by Garrett Baker (D0TheMath) · 2024-04-13T01:55:21.051Z · LW(p) · GW(p)

Is this coming from deep knowledge about Symbolica's method, or just on outside view considerations like "usually people trying to think too big-brained end up failing when it comes to AI".

Replies from: faul_sname
comment by faul_sname · 2024-04-13T03:53:17.557Z · LW(p) · GW(p)

Outside view (bitter lesson).

Or at least that's approximately true. I'll have a post on why I expect the bitter lesson to hold eventually, but is likely to be a while. If you read this blog post you can probably predict my reasoning for why I expect "learn only clean composable abstraction where the boundaries cut reality at the joints" to break down as an approach.

Replies from: D0TheMath
comment by Garrett Baker (D0TheMath) · 2024-04-13T16:33:19.019Z · LW(p) · GW(p)

I don’t think the bitter lesson strictly applies here. Since they’re doing learning, and the bitter lesson says “learning and search is all that is good”, I think they’re in the clear, as long as what they do is compute scalable.

(this is different from saying there aren’t other reasons an ignorant person (a word I like more than outside view in this context since it doesn’t hide the lack of knowledge) may use to conclude they won’t succeed)

Replies from: faul_sname
comment by faul_sname · 2024-04-13T18:40:54.558Z · LW(p) · GW(p)

By building models which reason inductively, we tackle complex formal language tasks with immense commercial value: code synthesis and theorem proving.

There are commercially valuable uses for tools for code synthesis and theorem proving. But structured approaches of that flavor don't have a great track record of e.g. doing classification tasks where the boundary conditions are messy and chaotic, and similarly for a bunch of other tasks where gradient-descent-lol-stack-more-layer-ML shines.

comment by sapphire (deluks917) · 2024-04-08T10:19:29.456Z · LW(p) · GW(p)

Does anyone have high quality analysis of how effective machines are for strength training and building muscles. Not free weights specifically machines. Im not the pickiest on how one operationalizes 'work'. More interested in the quality of the analysis. But some questions:

-- Do users get hurt frequently? Are the injuries chronic? (This is the most important question)

-- Do people who use them consistently gain muscle

-- Can you gain a 'lot' of muscle and strength liek you can with free weights. Or people cap out quickly if they are fit

-- Does strength from free weight carry over to other things. To barbells? To bodyweight calisthenics? To sports?

As I said im not the pickiest. Id rather have a great analysis of 'carryover to curling' than a meh quality analysis that covered all of my questions.

comment by ProgramCrafter (programcrafter) · 2024-03-23T19:48:24.254Z · LW(p) · GW(p)

I've came across a poll about exchanging probability estimates with another rationalist: https://manifold.markets/1941159478/you-think-something-is-30-likely-bu?r=QW5U.

You think something is 30% likely but a friend thinks 70%. To what does that change your opinion?

I feel like there can be specially-constructed problems when the result probability is 0, but haven't been able to construct an example. Are there any?

Replies from: thomas-kwa, Throwaway2367
comment by Thomas Kwa (thomas-kwa) · 2024-03-24T18:06:14.635Z · LW(p) · GW(p)

There is a box which contains money iff the front and back are painted the same color. Each side is independently 30% to be blue, and 70% to be red. You observe that the front is blue, and your friend observes that the back is red.

comment by Throwaway2367 · 2024-03-24T12:50:48.706Z · LW(p) · GW(p)

"No one assigns 70% to this statement." (Yes, your friend is an idiot, but that can be remedied, if needed, with a slight modification in the statement)

comment by niplav · 2024-03-19T20:14:16.260Z · LW(p) · GW(p)

Does anyone recommend a good book on finance? I'd like to understand most of the terms used in this post [LW · GW], and then some, but the old repository [LW · GW] doesn't have any suggestions.

Replies from: jmh, CstineSublime
comment by jmh · 2024-03-20T14:32:14.943Z · LW(p) · GW(p)

You might find this link helpful for your questions. 

This is a link to the glossory from the above site.

This is from the FRB of St. Louis.

Last, I would suggest you can also just ask any of the available LLM's out there now to explain the term you are interested in and get a pretty good initial explanation.

As for books, I have three. How good they are is subjective as one textbook is from years ago but they should cover most of the investment markets side of things:

Options as a Strategic Investment (Lawrence McMIllan)

Technical Analysis (Kirkpatrick & Dahlquist)

Investment Analysis and Portfolio Management (Frank Reilly) -- the old textbook I kept around.

If your interest in more in the economic terms and theory area you might look for The MIT Dictionary of Modern Economics or a similar dictionary of economic terms.

Replies from: niplav
comment by niplav · 2024-03-20T16:03:41.329Z · LW(p) · GW(p)

Thanks! I know about Investopedia & often use GPT-4 for this kind of thing, but I prefer books because I can also read them offline.

comment by CstineSublime · 2024-03-20T01:27:40.761Z · LW(p) · GW(p)

I don't see many particularly exotic finance terms, but I would think a recent edition of Brigham and Ehrhardt's "Financial Management: Theory and Practice" is probably the most authoritative (but not being available for 'preview' on Google Books I haven't been able to do a quick ctrl+f to see if it uses a sample of terms in that post). However I suspect that even one of those Tony Robbins books on investing will provide you the terminology or even Investopedia.

comment by EStokes · 2024-04-17T07:42:21.649Z · LW(p) · GW(p)

Hi any it may concern,

You could say I have a technical moat in a certain area and came across an idea/cluster of ideas that seemed unusually connected and potentially alignment-significant but whose publication seems potentially capabilities-enhancing. (I consulted with one other person and they also found it difficult to ascertain or summarize)

I was considering writing to EY on here as an obvious person who would both be someone more likely to be able to determine plausibility/risk across a less familiar domain and have an idea of what further to do. Is there any precedent or better idea for my situation? I suppose a general version is: "How can someone concerned with risk who can't easily further ascertain things themselves determine whether something is relevant to alignment/capabilities without increasing risk?" 

feel free to ask any questions that could help, and thank you anyone!

comment by Mateusz Bagiński (mateusz-baginski) · 2024-03-22T19:36:27.566Z · LW(p) · GW(p)

Bug report: I got notified about Joe Carlsmith's most recent post twice, the second time after ~4 hours

Replies from: joekc
comment by Joe Carlsmith (joekc) · 2024-03-22T19:49:16.401Z · LW(p) · GW(p)

That post ran into some cross-posting problems so had to re-do

comment by Mateusz Bagiński (mateusz-baginski) · 2024-03-17T04:19:43.090Z · LW(p) · GW(p)

Has anybody tried to estimate how prevalent sexual abuse is in EA circles/orgs compared to general population?

Replies from: habryka4
comment by habryka (habryka4) · 2024-03-17T05:58:03.024Z · LW(p) · GW(p)

It's really hard to get any kind of baseline here, and my guess is it differs hugely between different populations, but my guess (based on doing informal fermis here a bunch of times over the years) would be a lot lower than the average for the population, at least because of demographic factors, and then probably some extra.

Replies from: Nathan Young
comment by Nathan Young · 2024-03-19T02:01:55.744Z · LW(p) · GW(p)

Yeah and a couple of relevant things:

  1. The time EA sexual abuse article includes 1 person who isn't an EA and a sort of vibe that iirc includes most tech houses in the bay in the heading of "EA". This is inaccurate.
  2. EA takes a pretty strong stance on sexual harassment. Look at what people are banned from the forum for and scale it up. I've heard about people being banned from events for periods of time for causing serial discomfort. Compare this to Church communities I've been a part of and political communities and this is much stricter.