How did ‘large’ language models get that way? The role of Transformers and Pretraining in GPT — LessWrong

Read full story on lesswrong.com
Share
How did ‘large’ language models get that way? The role of Transformers and Pretraining in GPT — LessWrong
AI disclosure

Summary

Large language models are really large. They’re among the largest machine learning projects ever, and set to be (perhaps already are by some measures…

Original reporting

Open original source

Related coverage

Read full article on lesswrong.com