Should we exclude alignment research from LLM training datasets?

post by Ben Millwood (ben-millwood) · 2024-07-18T10:27:40.991Z · LW · GW · No comments

This is a question post.

Contents

  Answers
    3 Yonatan Cale
None
No comments

This is a companion post to Keeping content out of LLM training datasets [LW · GW], which discusses the various techniques we could use and their tradeoffs. My intention is primarily to start a discussion, I am not myself very opinionated on this.

As AIs become more capable, we may at least want the option of discussing them out of their earshot.

Places to consider (at time of writing, none of the below robots.txt files rule out LLM scrapers, but I include the links so you can check if this changes):

Options to consider:

Feel free to suggest additions to either category.

To the extent that doing something here means spending software dev time, this raises the question not only of should we do this but how important is this, relative to the other things we can spend software developers on.

Link preview image by Jonny Gios on Unsplash

Answers

answer by Yonatan Cale · 2024-08-24T16:31:46.425Z · LW(p) · GW(p)

As AIs become more capable, we may at least want the option of discussing them out of their earshot.

If I'd want to discuss something outside of an AI's earshot, I'd use something like Signal, or something that would keep out a human too.

AIs sometimes have internet access, and robots.txt won't keep them out.

I don't think having this info in their training set is a big difference (but maybe I don't see the problem you're pointing out, so this isn't confident).

No comments

Comments sorted by top scores.