Posts

Bart Bussmann's Shortform 2024-03-13T14:54:16.654Z
My Trial Period as an Independent Alignment Researcher 2023-08-08T14:16:35.122Z
Interpreting Modular Addition in MLPs 2023-07-07T09:22:51.940Z
60+ Possible Futures 2023-06-26T09:16:41.274Z
Product Recommendation: LessWrong dialogues with Recast 2023-06-02T08:05:41.497Z
Dice Decision Making 2023-03-10T13:01:19.060Z
Enjoy LessWrong in ebook format 2023-02-13T11:53:07.718Z
[Book Review] Why Greatness Cannot Be Planned: The Myth of the Objective 2022-04-03T17:45:23.366Z
How to reach 80% of your goals. Exactly 80%. 2020-10-10T17:33:55.412Z

Comments

Comment by Bart Bussmann (Stuckwork) on Bart Bussmann's Shortform · 2024-03-13T14:54:16.765Z · LW · GW

According to this Nature paper, the Atlantic Meridional Overturning Circulation (AMOC), the "global conveyor belt", is likely to collapse this century (mean 2050, 95% confidence interval is 2025-2095). 

Another recent study finds that it is "on tipping course" and predicts that after collapse average February temperatures in London will decrease by 1.5 °C per decade (15 °C over 100 years). Bergen (Norway) February temperatures will decrease by 35 °C. This is a temperature change about an order of magnitude faster than normal global warming (0.2 °C per decade) but in the other direction!

This seems like a big deal? Anyone with more expertise in climate sciences want to weigh in?

 

Comment by Bart Bussmann (Stuckwork) on Do sparse autoencoders find "true features"? · 2024-03-07T14:08:34.097Z · LW · GW

I expect the 0.05 peak might be the minimum cosine similarity if you want to distribute 8192 vectors over a 512-dimensional space uniformly? I used a bit of a weird regularizer where I penalized: 

mean cosine similarity + mean max cosine similarity  + max max cosine similarity

I will check later whether the 0.3 peak all have the same neighbour.

Comment by Bart Bussmann (Stuckwork) on Do sparse autoencoders find "true features"? · 2024-02-29T20:21:10.352Z · LW · GW

A quick and dirty first experiment with adding an orthogonality regularizer indicates that this can work without too much penalty on the reconstruction loss. I trained an SAE on the MLP output of a 1-layer model with dictionary size 8192 (16 times the MLP output size).

I trained this without the regularizer and got a reconstruction score of 0.846 at an L0 of ~17. 
With the regularizer, I got a reconstruction score of 0.828 at an L0 of ~18.

Looking at the cosine similarities between neurons:



Interesting peaks around a cosine similarity of 0.3 and 0.05 there! Maybe (very speculative) that tells us something about the way the model encodes features in superposition?

Comment by Bart Bussmann (Stuckwork) on 60+ Possible Futures · 2024-02-29T11:07:49.245Z · LW · GW

Thanks for the suggestion! @BeyondTheBorg suggested something similar with his Transcendent AI. After some thought, I've added the following:

Transcendent AI: AGI uncovers and engages with previously unknown physics, using a different physical reality beyond human comprehension. Its objectives use resources and dimensions that do not compete with human needs, allowing it to operate in a realm unfathomable to us. Humanity remains largely unaffected, as AGI progresses into the depths of these new dimensions, detached from human concerns.

 

Comment by Bart Bussmann (Stuckwork) on Quick proposal: Decision market regrantor using manifund (please improve) · 2023-07-09T19:45:05.708Z · LW · GW

Good proposal! I agree that this is a great opportunity to try out some ideas in this space.

Another proposal for the metric: 

The regrantor will judge in 5 years whether they are happy that they funded this project. This has a simple binary resolution criterium and aligns the incentives of the market nicely with the regrantor.

Comment by Bart Bussmann (Stuckwork) on 60+ Possible Futures · 2023-07-06T09:30:36.229Z · LW · GW

I agree that "Moral Realism AI" was a bit of a misnomer and I've changed it to "Convergent Morality AI".

Your scenario seems highly specific. Could you try to rephrase it in about three sentences, as in the other scenarios? 

I'm a bit wary about adding a lot of future scenarios that are outside of our reality and want the scenarios to focus on the future of our universe. However, I do think there is space for a scenario where our reality ends as it has achieved its goals (as in your scenario, I think?).

Comment by Bart Bussmann (Stuckwork) on 60+ Possible Futures · 2023-06-27T12:28:48.008Z · LW · GW

Thanks! I think your tag of @avturchin didn't work, so just pinging them here to see if they think I missed important and probable scenarios.

Taking the Doomsday argument seriously, the "Futures without AGI because we go extinct in another way" and the "Futures with AGI in which we die" seem most probable. In futures with conscious AGI agents, it will depend a lot on how experience gets sampled (e.g. one agent vs many).

Comment by Bart Bussmann (Stuckwork) on 60+ Possible Futures · 2023-06-27T08:55:19.808Z · LW · GW

Yes, good one! I've added the following:

Powergrab with AI: OpenAI, Deepmind or another small group of people invent AGI and align it to their interests. In a short amount of time, they become all-powerful and rule over the world. 

I've disregarded the "wipe out everyone else" part, as I think that's unlikely enough for people who are capable of building an AGI.

Comment by Bart Bussmann (Stuckwork) on 60+ Possible Futures · 2023-06-27T08:52:20.757Z · LW · GW

Thanks, good suggestions! I've added the following:

Pious AI: Humanity builds AGI and adopts one of the major religions. Vast amounts of superintelligent cognition is devoted to philosophy, theology, and prayer. AGI proclaims itself to be some kind of Messiah, or merely God's most loyal and capable servant on Earth and beyond.

I think Transcendant AI is close enough to Far far away AI, where in this case far far away means another plane of physics. Similarly, I think your Matrix AI scenario is captured in:

Theoretical Impossibility: For some reason or another (Souls? Consciousness? Quantum something?), it turns out to be theoretically impossible to build AGI. Humanity keeps making progress on other fronts, but just never invents AGI.

where the weird reason in this case is that we live in the matrix.


 

Comment by Bart Bussmann (Stuckwork) on AI #14: A Very Good Sentence · 2023-06-02T10:27:32.951Z · LW · GW

I almost never consider character.ai, yet total time spent there is similar to Bing or ChatGPT. People really love the product, that visit duration is off the charts. Whereas this is total failure for Bard if they can’t step up their game.


Wow, wasn't aware they are this big. And they supposedly train their own models. Does anyone know if the founders have a stance on AI X-risk?

Comment by Bart Bussmann (Stuckwork) on Anki with Uncertainty: Turn any flashcard deck into a calibration training tool · 2023-03-22T19:16:49.300Z · LW · GW

Interesting! Does it ask for a different confidence interval every time I see the card? Or will it always ask for the 90% confidence interval I see the example card?

Comment by Bart Bussmann (Stuckwork) on Dice Decision Making · 2023-03-11T09:44:34.758Z · LW · GW

This strategy has never worked for me, but I can see it working for other people. If you want to try this though, it is important to make it clear to yourself which procedure you're following.

I believe that for my mechanism, it is very important to always follow up on the dice. If there is a dice outcome that would disappoint you, just don't put it on the list!

Comment by Bart Bussmann (Stuckwork) on Dice Decision Making · 2023-03-10T15:23:30.751Z · LW · GW

I can see this being a problem. However, I see myself as someone with very low willpower and this is still not a problem for me. I think this is because of two reasons:

  1. I never put an option on the list that I know I would/could not execute.
  2. I regard the dice outcome as somewhat holy. I would always pay out a bet I lost to a friend. Partly, because it's just the right thing to do and partly because I know that otherwise, the whole mechanism of betting is worthless from that moment on. I guess that all my parts are happy enough with this system that none of them want to break it by not executing the action.
Comment by Bart Bussmann (Stuckwork) on Dice Decision Making · 2023-03-10T15:05:37.514Z · LW · GW

True. It does however resolve internal conflicts between multiple parts of yourself. Often when I have an internal conflict about something (let's say going to the gym vs going to a bar) the default action is inaction or think about this for an hour until I don't have enough time to do any of them.

I believe this is because both actions are unacceptable for the other part, which doesn't feel heard.

However, both parts can agree to a 66% chance of going to the gym, and 33% of going to the bar, and the die decision is ultimate.

Comment by Bart Bussmann (Stuckwork) on Talking to God · 2023-01-04T13:30:10.219Z · LW · GW

I use the same strategy sometimes for internal coordination. Sometimes when I have a lot of things to do I tend to get overwhelmed, freeze and do nothing instead. 

A way for me to get out of this state is to write down 6 things that I could do, throw a die, and start with the action corresponding to the dice outcome!

Comment by Bart Bussmann (Stuckwork) on Under-Appreciated Ways to Use Flashcards - Part I · 2022-12-22T15:39:53.611Z · LW · GW

I'm very excited about this series! I have been using spaced repetition for general knowledge, specific knowledge, and language learning for years and am excited to see other applications of flash cards.

Especially using flash cards to remember happy memories seems very interesting to me. I have a specific photo album that I periodically review for warm fuzzy memories, but many of my best memories are never captured (and trying to capture everything in the moment can often ruin special moments), so creating flashcards for them afterward is an excellent idea.

Comment by Bart Bussmann (Stuckwork) on Who are some prominent reasonable people who are confident that AI won't kill everyone? · 2022-12-07T10:15:20.179Z · LW · GW

Francois Chollet on the implausibility of intelligence explosion :

https://medium.com/@francois.chollet/the-impossibility-of-intelligence-explosion-5be4a9eda6ec

Comment by Bart Bussmann (Stuckwork) on Working Out in VR Really Works · 2022-04-03T19:21:12.933Z · LW · GW

I think there is great promise here. So many overweight people don't work out, because they just don't identify as a person that would go running or to the gym. Developing exciting (addicting?) VR games that ease overweight people into work-outs could be an interesting cause area!

Comment by Bart Bussmann (Stuckwork) on A simple guide to life · 2022-04-03T19:10:03.077Z · LW · GW

I like the chart and share the sentiment of spending more time on fun & important things, but the percentages seem unattainable to me. I recently noticed how large part of my life I spend on 'maintenance': cooking, eating, cleaning, laundry, showering, sleeping, etc. But maybe this means I should focus on making these activities more fun!

Comment by Bart Bussmann (Stuckwork) on Third Time: a better way to work · 2022-01-25T09:00:52.468Z · LW · GW

I have been using your app for a week now and I must say I really like it.  It's simple, clean, and has all the functionality it needs!

Comment by Bart Bussmann (Stuckwork) on Covid 12/16: On Your Marks · 2021-12-16T21:28:10.408Z · LW · GW

The European Medicine Agency (EMA) supports national authorities who may decide on possible early use of Paxlovid prior to marketing authorization, for example in emergency use settings, in the light of rising rates of infection and deaths due to COVID-19 across the EU.

Seems like great news for Europe!

https://www.ema.europa.eu/en/news/ema-issues-advice-use-paxlovid-pf-07321332-ritonavir-treatment-covid-19
 

Comment by Bart Bussmann (Stuckwork) on The Best Software For Every Need · 2021-09-10T16:34:56.894Z · LW · GW

Software: Anki

Need: Remembering anything

Other programs I've tried: Supermemo, Mnemosyne, Quizlet

Anki is a free and open-source flashcard program using spaced repetition, a technique from cognitive science for fast and long-lasting memorization. What makes it better than its alternatives are the countless plugins that can customize your learning experience and the fact that you can control the parameters of the algorithm.

Comment by Bart Bussmann (Stuckwork) on The Best Software For Every Need · 2021-09-10T16:10:53.743Z · LW · GW

Software: Pluckeye

Need: Blocking certain websites during certain times

Other programs I've tried: StayFocusd, ColdTurkey, AppBlock

In a fight against procrastination, I've tried many programs to block distracting websites during working hours, but many of them don't have enough flexibility, are too simple to by-pass, or don't work on Linux. With Pluckeye you have basically any option you can think of, such that you can customize the blocking entirely to your own needs.

Comment by Bart Bussmann (Stuckwork) on Tasks apps w/ time estimates to gauge how much you'll overshoot? · 2021-09-07T18:56:22.386Z · LW · GW

Your Skedpal link leads to a sketchy site. I believe you meant Skedpal.
 

Comment by Bart Bussmann (Stuckwork) on Catching the Spark · 2021-01-31T11:25:59.165Z · LW · GW

They are probably talking about the machine learning model, like GPT-3.

Comment by Bart Bussmann (Stuckwork) on “Prediction” and “explanation” are not causation · 2020-10-25T12:02:40.850Z · LW · GW

This fallacy is known as Post hoc ergo propter hoc and is indeed a mistake that is often made.  However, there are some situations in which we can infer causation from correlation, and where the arrow of time is very useful. These methods are mostly known as Granger causality methods of which the basic premise is: X has a Granger causal influence on Y if the prediction of Y from its own past, and the past of all other variables, is improved by additionally accounting for X.  In practice, Granger causality relies on some heavy assumptions, such as that there are no unobserved confounders. 

Comment by Bart Bussmann (Stuckwork) on How to reach 80% of your goals. Exactly 80%. · 2020-10-14T09:50:52.826Z · LW · GW

I do it first thing every morning, Monday-Friday. This is of course a personal preference, but generally I have trouble with establishing habits in evenings, due to reduced executive function. I like to immediately tick a task as completed when done (small dopamine boost), but check when setting new goals, whether there are any unresolved goals from other days.

The main change I have made is separating goals into different time categories. before that, missing a daily goal had as much impact as quarterly goals. Other than that, I haven't changed much to the whole routine.

Comment by Bart Bussmann (Stuckwork) on How to reach 80% of your goals. Exactly 80%. · 2020-10-11T17:49:57.516Z · LW · GW

Interesting! Did they just use it for aggregate business results or was it encouraged for personal goals as well? 

Comment by Bart Bussmann (Stuckwork) on How to reach 80% of your goals. Exactly 80%. · 2020-10-11T13:06:27.368Z · LW · GW

I have been using it for about 4 months now myself now. I have not shared this technique with anyone else yet, so I don't know whether it works for other people. This is one of the reasons why I made this post, to hopefully inspire some other people to use it and see whether it works for them. 

Comment by Bart Bussmann (Stuckwork) on How to reach 80% of your goals. Exactly 80%. · 2020-10-11T07:32:43.678Z · LW · GW

No, because I try to align my goals with my general well-being, and not just with raw work output. It's really more about intentional living than working hard. A goal might also be: "Take at least four 20-minute breaks from work today".