LessWrong 2.0 Reader
View: New · Old · Top← previous page (newer posts) · next page (older posts) →
Kendra · 2013-10-01T08:27:40.475Z · comments (0)
Lumina is incredibly cheap right now. I pre-ordered for 250usd. Even genuinely quite poor people I know don't find the price off-putting (poor in the sense of absolutely poor for the country they live in). I have never met a single person who decided not to try Lumina because the price was high. If they pass its always because they think its risky.
silentbob on The Alignment Problem No One Is Talking AboutJust to note your last paragraph reminds me of Stuart Russel's approach to AI alignment in Human Compatible. And I agree this sounds like a reasonable starting point.
jett on Transformers Represent Belief State Geometry in their Residual StreamThis is such a cool result! I tried to reproduce it in this notebook
In terms of the big labs being inefficient, with hindsight perhaps. Anyway I have said that I can't understand why they aren't putting much more effort into Dishbrain etc. If I had ~$1B and wanted to get ahead on a 5 year timescale I would give it more probability expectation etc.
For
I’m probably typical-minding a bit here, but: you say you have had mental health issues in the past (which, based on how you describe them, sound at least superficially similar to my own), and that you feel like you’ve outlived yourself. Which, although it is a feeling I recognise, is still a surprising thing to say: even a high P(doom) only tells you that your life might soon have to stop, not that it already has! My wild-ass guess would be that, in addition to maybe having something to prove intellectually and psychologically, you feel lost, with the ability to do things (btw, I didn’t know your blog and it’s pretty neat) but nothing in particular to do. Maybe you’re considering finishing your degree because it gives you a medium-term goal with some structure in the tasks associated with it?
aaron-bergman on quila's ShortformThank you, that is all very kind! ☺️☺️☺️
I expect if he continues being what he is, he'll produce lots of cool stuff which I'll learn from later.
I hope so haha
jett on Transformers Represent Belief State Geometry in their Residual StreamFor the two sets of mess3 parameters I checked the stationary distribution was uniform.
ben-lang on Losing Faith In ContrarianismNice post. Gets at something real.
My feeling is that a lot of contrarians get "pulled into" a more contrarian view. I have noticed myself in discussions propose a (specific, technical point correcting a detail of a particular model). Then, when I talk to people about it I feel like they are trying to pull me towards the simpler position (all those idiots are wrong, its completely different from that). This happens with things like "ah, so you mean...", which is very direct. But also through a much more subtle process, where I talk to many people, and most of them go away thinking "Ok, specific technical correction on a topic I don't care about that much." and most of them never talk or think about it again. But the people who get the exaggerated idea are more likely to remember.
russellthor on Against "argument from overhang risk"If you are referring to this:
If we institute a pause, we should expect to see (counterfactually) reduced R&D investment in improving hardware capabilities, reduced investment in scaling hardware production, reduced hardware production, reduced investment in research, reduced investment in supporting infrastructure, and fewer people entering the field.
This seems an extreme claim to me (if these effects are argued to be meaningful), especially "fewer people entering the field"! Just how long do you think you would need a pause to make fewer people enter the field? I would expect that not only would the pause have to have lasted say 5+ years but there would have to be a worldwide expectation that it would go on for longer to actually put people off.
Because of flow on effects and existing commitments, reduced hardware R&D investment wouldn't start for a few years either. Its not clear that it will meaningfully happen at all if we want to deploy existing LLM everywhere also. For example in robotics I expect there will be substantial demand for hardware even without AI advances as our current capabilities havn't been deployed there yet.
As I have said here, and probably in other places, I am quite a bit more in favor of directly going for a hardware pause specifically for the most advanced hardware. I think it is achievable, impactful, and with clearer positive consequences (and not unintended negative ones) than targeting training runs of an architecture that already seems to be showing diminishing returns.
If you must go for after FLOPS for training, then build in large factors of safety for architectures/systems that are substantially different from what is currently done. I am not worried about unlimited FLOPS on GPT-X but could be for >100* less on something that clearly looks like it has very different scaling laws.
connor-kissane on Sparse Autoencoders Work on Attention Layer OutputsThanks for the comment! We always use the pre-ReLU feature activation, which is equal to the post-ReLU activation (given that the feature is activate), and is purely linear function of z. Edited the post for clarity.