

(And the essence of what I’ll say applies just as well to other current “large language models” as to ChatGPT.) I should say at the outset that I’m going to focus on the big picture of what’s going on-and while I’ll mention some engineering details, I won’t get deeply into them. But how does it do it? And why does it work? My purpose here is to give a rough outline of what’s going on inside ChatGPT-and then to explore why it is that it can do so well in producing what we might consider to be meaningful text.

That ChatGPT can automatically generate something that reads even superficially like human-written text is remarkable, and unexpected. “Wolfram|Alpha as the Way to Bring Computational Knowledge Superpowers to ChatGPT” » A discussion about the history of neural nets »
