Quick Thoughts on Language Models

post by RohanS · 2023-07-18T20:38:57.866Z · LW · GW · 0 comments

Epistemic status: Somewhere between Exploratory and My Best Guess. Plenty of other people have written about similar or related ideas, and about alternative or conflicting views.

Epistemic effort: Ideas developed slowly over about 3 years, via learning ML, playing with language models, reading work by others, and discussing language model capabilities with various people. I've had a few recent discussions particularly related to the ideas below. About 1 hour of writing and editing this before posting.

Thoughtful feedback on this post is very welcome!

Imagine somebody telling you to make up random words, and you say, "Morvelkainen bloombla ringa mongo."

Imagine a mind of a level - where, to be clear, I'm not saying GPTs are at this level yet -

Imagine a Mind of a level where it can hear you say 'morvelkainen blaambla ringa', and maybe also read your entire social media history, and then manage to assign 20% probability that your next utterance is 'mongo'.

The fact that this Mind could double as a really good actor playing your character, does not mean They are only exactly as smart as you.

When you're trying to be human-equivalent at writing text, you can just make up whatever output, and it's now a human output because you're human and you chose to output that.

GPT-4 is being asked to predict all that stuff you're making up. It doesn't get to make up whatever. It is being asked to model what you were thinking - the thoughts in your mind whose shadow is your text output - so as to assign as much probability as possible to your true next word.

0 comments

Comments sorted by top scores.