Posts
Comments
Quick update! Due to the pub being too full/loud, we'll be moving the meetup to Ezra and Haydn's place in ~15(?) minutes. See specific address on the Montreal Rationalists Discord.
Thanks for noticing and pointing it out!
The API token was cancelled, sorry about that. The most recent version of the chatbot is now at https://chat.stampy.ai and https://chat.aisafety.info, and should not have the API token issue.
I'm glad to hear you're trying to catch up with the alignment ecosystem!
It is still supposed to be live and active, and it still works for me. Are you sure you have https://alignmentsearch.up.railway.app? If so, then I'm not sure what's going on, it worked for everyone who I know that tried. If you have a different link, maybe we've been linking to the website incorrectly somewhere so please share the link you do have.
Edit: just realized you weren't speaking of https://alignmentsearch.up.railway.app, I thought it was a standalone comment. I'm getting the same 404 error for the aisafety.world link.
Thanks!
- Just did. :)
- We've seen the channel, yes, though haven't messaged in it or anything.
Thanks for the comment!
At this point, we don't have a very clear plan, other than thinking of functionalities and adding them as fast as possible in an order that seems sensible. The functionalities we want to add include:
- Automatic update of the dataset relatively often.
- Stream completions.
- Test embeddings using SentenceTransformers + Finetuning instead of OpenAI for cost and quality, and store them in Pinecone/Weaviate/Other (tbd); this will enable us to use the whole dataset for semantic search, and for the semantic similarity to have more 'knowledge' about technical terms used in the alignment space, which I expect to produce better results. We also want to test and add biases to favor 'good' sources to maximize the quality of semantic search. It's also possible that we'll make a smaller, more specialized dataset of curated content.
- Add modes and options. HyDE, Debate, Comment, Synthesis, temp, etc. Possibly add options to make use of GPT-4, depending on feasibility.
- Figure out how to make this scale without going bankrupt.
- Add thumbs-up/down for A/B testing prompt, the bias terms, and curated vs uncurated datasets.
- Add recommended next questions the user can ask, possibly taken from a question database.
- Improve UX/UI.
We have not taken much time (we were very pressed for it!) to consider the best way to onboard collaborators. We are communicating on our club's Discord server at the moment, and would be happy to add people who want to contribute, especially if you have experience in any of the above. DM me on Discord at BionicD0LPH1N#5326 or on LW.
The current version has trouble answering basic AI safety questions.
That's true sometimes, and a problem. We observe fewer such errors on the full dataset, and are currently working on having that up. Additional modes, like HyDE, and the bias mentioned earlier, might further improve results. Getting better embeddings + finetuning them on our dataset might improve search. Finally, when the thumbs up/down feature is up, we will be able to quickly search over a list of possible prompts we think might be more successful, and find the ones that reduce bad answers. Overall, I think that this is a very solvable problem, and are making rapid progress.
About curating the dataset (or favoring some types of content), we agree and are currently investigating the best ways to do this.
About walking people through the extended alignment bingo, this is a feature we're planning to add. Something that might make sense is to have a slider for 'level-of-expertise', where beginners would have more detailed answers that assume less knowledge, and get recommended further questions that guide them through the bad takes bingo.
The feedback function for wrong answers is one of our top priorities, and in the meantime we ask you give the failing question-answer pairs in our form.