alexlyzhov's Shortform
post by alexlyzhov · 2022-06-12T22:47:51.149Z · LW · GW · 6 commentsContents
6 comments
6 comments
Comments sorted by top scores.
comment by alexlyzhov · 2022-06-12T22:47:51.482Z · LW(p) · GW(p)
Every other day I have a bunch of random questions related to AI safety research pop up but I'm not sure where to ask them. Can you recommend any place where I can send these questions and consistently get at least half of them answered or discussed by people who are also thinking about it a lot? Sort of like an AI safety StackExchange (except there's no such thing), or a high-volume chat/discord. I initially thought about LW shortform submissions, but it doesn't really look like people are using the shortform for asking questions at all.
Replies from: przemyslaw-czechowski, niplav, harfe↑ comment by Jan Czechowski (przemyslaw-czechowski) · 2022-06-13T21:23:26.840Z · LW(p) · GW(p)
There's an AI safety camp slack with #no-stupid-questions channel. I think people stay there even after the camp ends (I'm still there although this year edition ended last week). So you can either apply for next years edition (which I very much recommend!) or maybe contact organizers if they can add you without you being AISC participant/alumni? Just a disclaimer, I'm not sure how active this slack is between camps, and it might be that lot of people leave after the camp ends.
↑ comment by niplav · 2022-06-13T18:39:10.231Z · LW(p) · GW(p)
The closest thing to an AI safety StackExchange is the stampy wiki, with loads of asked & answered questions. It also has a discord.
↑ comment by harfe · 2022-06-13T00:06:56.272Z · LW(p) · GW(p)
The eleuther.ai discord has two alignment channels with reasonable volume (#alignment-general and #alignment-beginners). These might be suitable for your needs.
Replies from: przemyslaw-czechowski↑ comment by Jan Czechowski (przemyslaw-czechowski) · 2022-06-13T21:28:57.570Z · LW(p) · GW(p)
I just logged in there, thanks - however they have a couple of disclaimers that this is not a place for beginner / technical support questions. But I suppose it's more to avoid people asking "how do I install pytorch?" rather than AI safety questions, so maybe it will suit the needs of OP
comment by alexlyzhov · 2023-05-11T19:38:38.607Z · LW(p) · GW(p)
Random idea about preventing model stealing. After finetuning a mixture of experts model with your magic sauce, place the trained experts on geographically distinct servers with heterogeneous tech stacks and security systems to avoid common vulnerabilities. Horcrux vibes