Review of 'GPT-3: Language Models are Few-Shot Learners (Paper Explained) - YouTube'

1 · Shane Mulligan · May 29, 2020, 4 p.m.
Work in progress Original video GPT-3: Language Models are Few-Shot Learners (Paper Explained) - YouTube Demonstration of me reading this video and taking notes GPT-3 GPT 3 has 175 billion parameters which this is absolutely crazy is an order of magnitude higher than anything that ever existed. GPT-2 parameters by comparison This is what people are talking about when they say parameters. 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 Parameters Single Transformer block Conv1d attn/c_a...