Posts

Comments

Comment by acertain (fread2281) on Alexander Gietelink Oldenziel's Shortform · 2024-11-01T07:13:20.820Z · LW · GW

I guess this is sorta about your 3, which I disbelieve (though algorithms for tasks other than learning are also important). Currently, Bayesian inference vs SGD is a question of how much data you have (where SGD wins except for very little data). For small to medium amounts of data, even without AGI, I expect SGD to lose eventually due to better inference algorithms. For many problems I have the intuition that it's ~always possible to improve performance with more complicated algorithms (eg sat solvers). All that together makes me expect there to be inference algorithms that scale to very large amounts of data (that aren't going to be doing full Bayesian inference but rather some complicated approximation).

Comment by acertain (fread2281) on Where is the Learn Everything System? · 2024-09-29T08:09:58.407Z · LW · GW

Relevant: https://andymatuschak.org/hmwl/

Comment by acertain (fread2281) on Why is o1 so deceptive? · 2024-09-28T01:23:45.098Z · LW · GW

o1 CoT: The user is asking for more references about brownies. <Reasoning about what the references should look like> So, the assistant should list these references clearly, with proper formatting and descriptions, and provide actual or plausible links. Remember, the model cannot retrieve actual URLs, so should format plausible ones.

this might encourage it to make up links

Comment by acertain (fread2281) on Sparsify: A mechanistic interpretability research agenda · 2024-04-28T05:44:31.658Z · LW · GW

description of (network, dataset) for LLMs ?= model that takes as input index of prompt in dataset, then is equivalent to original model conditioned on that prompt

Comment by acertain (fread2281) on Why is capnometry biofeedback not more widely known? · 2023-12-23T17:40:31.137Z · LW · GW

There exist inexpensive real co2 sensors, e.g. https://www.sparkfun.com/products/22396 . Datasheet says only updates every 5 seconds & 60s response time "for achieving 63% of a respective step function", which I guess is what parent comment means by "They’ll likely be extremely slow".

Probably worth searching e.g. digikey for sensors with faster response time.

Comment by acertain (fread2281) on Alexander Gietelink Oldenziel's Shortform · 2023-10-25T20:39:00.449Z · LW · GW

What about specialized algorithms for problems (e.g. planning algorithms)?

Comment by acertain (fread2281) on RSPs are pauses done right · 2023-10-14T20:29:08.319Z · LW · GW

IANAL, but I think that this is currently impossible due to anti-trust regulations.

I don't know anything about anti-trust enforcement, but it seems to me that this might be a case where labs should do it anyways & delay hypothetical anti-trust enforcement by fighting in court.

Comment by acertain (fread2281) on Frame Control · 2022-01-18T04:49:33.447Z · LW · GW

blueiris's posts read to me as a combination of good concepts & poor quality attacks/attempts to defend leverage (or something?). Personally I'd mind the attacks more if they were more successful and/or less obvious I think? As-is they're annoying but don't seem very dangerous epistemically.

Comment by acertain (fread2281) on What would we do if alignment were futile? · 2021-12-22T01:49:27.799Z · LW · GW

Trying to reduce the amount of compute risks increasing hardware overhang once that compute is rebuilt. I think trying to slow down capabilities research (e.g. by getting a job at an AI lab and being obstructive) is probably better.

edit: meh idk. Whether or not this improves things depends on how much compute you can destroy & for how long, ml scaling, politics, etc etc. But the current world of "only big labs with lots of compute budget can achieve SOTA" (arguable, but possibly more true in the future) and less easy stuff to do to get better performance (scaling) both seem good.

Comment by acertain (fread2281) on Discussion with Eliezer Yudkowsky on AGI interventions · 2021-12-22T00:02:28.599Z · LW · GW

I personally think work on reduced precision inference (e.g. 4 bit!) is probably useful, as circuits should be easier to analyze than floats.

Comment by acertain (fread2281) on What posts on finance would your find helpful or interesting? · 2021-07-28T20:53:24.543Z · LW · GW

How to convert simple predictions/probability distributions (e.g. $stock will go down with x% probability at a date distributed around day Y an amount normally distributed around Z) into positions.

How much should the average person worry about tail risk? the average EA?

Less naive portfolio construction.

What tools from quantitative finance might be useful outside of finance: Econometrics & probabilistic modeling as used in finance (or as used 8 years ago or whatever)? Risk modeling?