Open & Welcome Thread - Aug/Sep 2022
post by Thomas · 2022-08-03T10:22:53.266Z · LW · GW · 32 commentsContents
32 comments
If it’s worth saying, but not worth its own post, here's a place to put it.
If you are new to LessWrong, here's the place to introduce yourself. Personal stories, anecdotes, or just general comments on how you found us and what you hope to get from the site and community are invited. This is also the place to discuss feature requests and other ideas you have for the site, if you don't want to write a full top-level post.
If you're new to the community, you can start reading the Highlights from the Sequences, a collection of posts about the core ideas of LessWrong.
If you want to explore the community more, I recommend reading the Library, [? · GW] checking recent Curated posts [? · GW], seeing if there are any meetups in your area [? · GW], and checking out the Getting Started [? · GW] section of the LessWrong FAQ [? · GW]. If you want to orient to the content on the site, you can also check out the new Concepts section [? · GW].
The Open Thread tag is here [? · GW]. The Open Thread sequence is here [? · GW].
32 comments
Comments sorted by top scores.
comment by Thomas · 2022-08-03T10:29:59.022Z · LW(p) · GW(p)
If and only if a month has exactly two characters in its Roman numeral, then it has less than 31 days. No exception.
Save your knuckles!
Replies from: Charlie Steiner, Viliam, conor-sullivan↑ comment by Charlie Steiner · 2022-08-05T23:55:49.543Z · LW(p) · GW(p)
There's a poem I know:
Thirty days has September, April, June, and November.
All the rest have thirty-one.
Except for February.
↑ comment by Lone Pine (conor-sullivan) · 2022-08-03T12:15:20.120Z · LW(p) · GW(p)
Also, only eat oysters in a month with an R in the name.
comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2022-08-10T18:16:13.628Z · LW(p) · GW(p)
Desired feature for LessWrong:
Ability to register bets that's is linked to accounts. Leaderboards, offering and accepting bets, ability to assign arbiters etc
It would be cool to see and track various bets - LessWrong is the obvious platform
The issue with forecasting platforms is that for any one of them there is a huge chance they won't be around irn 5 or 10 years which defeats the point of long term bets
EDIT: another feature we desperately need: right now if I'm logged in on the alignment forum it doesn't show me the lesswrong comments. That's okay & by design. However if would be good if it at least told me that there are LW comments on the alignmentForum post in the first place - that I may then decide to read or not read. It can be very confusing this way: finding a post without any comments that you could have sworn had more engagement last time you checked.
Replies from: Raemon↑ comment by Raemon · 2022-08-10T18:18:48.825Z · LW(p) · GW(p)
I've thought about integrating bets and predictions more into LessWrong. An alternate option is to integrate with existing prediction platforms to see people's track record on Metaculus, PredictionBook, etc.
Replies from: alexander-gietelink-oldenziel↑ comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2022-08-10T18:43:35.380Z · LW(p) · GW(p)
Wooh!
I'm curious what your thoughts about the various options are ?
comment by Adam Zerner (adamzerner) · 2022-09-19T03:34:17.077Z · LW(p) · GW(p)
In general, even in the rationality community, people's reactions, including my own, to the fact that doom seems imminent -- whether it's in 5 years or 50 years -- seems much too small. I wonder how much of this is because it feels science fiction-y.
If it was nuclear war, would that change things? An asteroid hitting? What about whether it is mainstream people vs non-mainstream people pulling the alarm? If a majority of mainstream academics were pulling the alarm on an asteroid hitting in the next 5-50 years, would reactions be different?
Replies from: jacob_cannell↑ comment by jacob_cannell · 2022-09-28T04:58:27.670Z · LW(p) · GW(p)
In the worlds where we have AI doom it's likely because of large amounts of easy optimization slack that AGI exploits leading to hard takeoff, or perhaps coordination failures and deceptive alignment in slower takeoff scenarios. Either way there doesn't seem to be much one can do about that other than contribute to AI safety.
Contrast to nuclear war, where more concrete conventional preparation like bomb shelters and disaster survival preparation has at least some non-epsilon expected payout.
Also, most of the current leaders/experts in AI don't put much probability on doom compared to LW folks.
comment by Rafael Harth (sil-ver) · 2022-09-13T09:27:46.039Z · LW(p) · GW(p)
Does anyone happen to know the maximum number of inputs to any element in a digital computer at the level of electronic gates? E.g., there are of course AND and OR gates, so the number is at least 2. But you can also set up an electronic circuit such that it implements, e.g., an AND with 3 inputs (though this is never required since you can make a 3-input AND out of two 2-input ANDs, and similarly for any other gate with 3+ inputs). Hence the question, what is the highest number of inputs that does physically exist in a typical CPU?
comment by niplav · 2022-08-11T22:17:29.222Z · LW(p) · GW(p)
I'm pretty proud of this and will take the option to brag a little :-)
I've found an integer sequence that is both meaningfully generateable and doesn't occur in the OEIS (and superseeker also doesn't find any way to generate it).
The first six integers in the sequence are 1,2,8,144,10752,3306240, any further ones are (for me at the moment) basically impossible to compute (they are divisible by the factorial numbers, which makes sense (in context), the sequence divided by the factorial numbers is 1,2,4,24,448,27552).
I'd be interested if anyone finds some interesting structure in the existing numbers, I've only noticed that they have surprisingly small prime factors.
comment by Fer32dwt34r3dfsz (rodeo_flagellum) · 2022-09-25T21:07:55.706Z · LW(p) · GW(p)
Has anyone here considered working on the following?:
https://www.super-linear.org/prize?recordId=recT1AQw4H7prmDE8
$500 prize pool for creating an accurate, comprehensive, and amusing visual map of the AGI Safety ecosystem, similar to XKCD’s map of online communities or Scott Alexander’s map of the rationalist community.
Payout will be $400 to the map which plex thinks is highest quality, $75 to second place, $25 to third. The competition will end one month after the first acceptable map is submitted, as judged by plex.
Resources, advice, conditions:
- This is a partial list of items which might make sense to include.
- You are advised to iterate, initially posting a low-effort sketch and getting feedback from others in your network, then plex.
- You may create sub-prizes on Bountied Rationality for the best improvements to your map (if you borrow ideas from other group’s public bounties and win this prize you must cover the costs of their bounty payouts).
- You may use DALL-E 2 or other image generation AIs to generate visual elements.
- You may collaborate with others and split the prize, but agree internally on roles and financial division, and distribute it yourselves.
- You can use logos as fair-use, under editorial/educational provision.
- You can scale items based on approximate employee count (when findable), Alexa rank (when available), number of followers (when available) or wild guess (otherwise).
- You agree to release the map for public use.
I asked about FLI's map in this question [LW · GW] and it received some traction. I might go ahead and try this, starting with FLI's map and expanding off of it.
comment by stuhlmueller · 2022-08-05T17:46:06.573Z · LW(p) · GW(p)
Is there a keyboard shortcut for “go to next unread comment” (i.e. next comment marked with green line)? In large threads I currently scroll a while until I find the next green comment, but there must be a better way.
comment by Q Home · 2022-09-17T13:18:12.391Z · LW(p) · GW(p)
Hello! I've heard [LW(p) · GW(p)] you can ask people about your content in the open thread. Sorry if I'm asking too soon.
Could you help me to explain this ("Should AI learn human values, human norms or something else?" [LW · GW]) idea better? It's a 3 minute read.
I also would like to discuss with somebody those thought experiments [LW · GW]. Not in a 100% formal way.
comment by Yoav Ravid · 2022-08-05T06:10:07.862Z · LW(p) · GW(p)
Is it just me or has the center part of the homepage moved slightly left? Something just feels off when I look at it.
comment by Mitchell_Porter · 2022-09-22T20:26:00.472Z · LW(p) · GW(p)
comment by Torello · 2022-09-17T17:56:02.466Z · LW(p) · GW(p)
My notes and comments on the book Noise:
https://digitalsauna.wordpress.com/2022/09/17/noise-by-kahneman-sibony-sunstein-2021/
comment by Alex Vermillion (tomcatfish) · 2022-08-29T03:11:01.956Z · LW(p) · GW(p)
Someone recently asked me if there was audio for https://www.lesswrong.com/highlights [? · GW] compiled into a playlist anywhere. I can't seem to find it; does this already exist?
Replies from: Raemon↑ comment by Raemon · 2022-08-29T03:21:01.312Z · LW(p) · GW(p)
As of last Friday, you can actually play the audio for each post directly from the post page. We may figure out a playlist in another format. I'm curious what your preferred format would be.
Replies from: tomcatfish↑ comment by Alex Vermillion (tomcatfish) · 2022-09-01T18:44:11.860Z · LW(p) · GW(p)
I'll pass this along! It wasn't for me, it was for someone who apparently likes to queue up LW posts in their podcast player and wanted to know if there was an automatic way to do so.
comment by Alex Vermillion (tomcatfish) · 2022-08-22T03:40:19.852Z · LW(p) · GW(p)
Month 2 of asking how to edit our group!!! [LW(p) · GW(p)]
Replies from: D0TheMath, Raemon↑ comment by Garrett Baker (D0TheMath) · 2022-08-22T04:38:22.405Z · LW(p) · GW(p)
Obvious solution: Have you tried using the Intercom to contact the team?
Replies from: habryka4, tomcatfish↑ comment by habryka (habryka4) · 2022-08-23T23:26:34.597Z · LW(p) · GW(p)
Yeah, Intercom is the right way to reach us, and we should usually respond within an hour or two.
↑ comment by Alex Vermillion (tomcatfish) · 2022-08-24T23:02:53.308Z · LW(p) · GW(p)
Thank you, I had not.
↑ comment by Raemon · 2022-08-24T00:03:22.616Z · LW(p) · GW(p)
(it looks like you have intercom hidden. I just un-hid it for you, so you can message us with it. Use the circle in the bottom-right corner)
Replies from: tomcatfish↑ comment by Alex Vermillion (tomcatfish) · 2022-08-29T03:09:55.571Z · LW(p) · GW(p)
Thank you for doing that; I'll do this some day soon when I'm ready to keep my laptop open on the tab for a while.
Replies from: Raemoncomment by Denver Greene (denver-greene) · 2022-08-04T04:56:38.425Z · LW(p) · GW(p)
Is there a way to get the book in a physical form? I want to share it with a family member who expressed interest, but they really need a physical book to read. I could just get the epub printed I'm sure, but I'd like to purchase it in a way that helps the organization if possible.
Replies from: ChristianKl↑ comment by ChristianKl · 2022-08-04T07:18:25.010Z · LW(p) · GW(p)
Which book do you mean?
Replies from: denver-greene↑ comment by Denver Greene (denver-greene) · 2022-08-05T17:49:20.373Z · LW(p) · GW(p)
I'm thinking Rationality AI to Zombies, or whatever Less Wrong has to provide.
Replies from: Hoagy