Stanford CS224N: NLP with Deep Learning | Winter 2019 | Lecture 20 – Future of NLP + Deep Learning
For more information about Stanford’s Artificial Intelligence professional and graduate programs, visit: https://stanford.io/3Cbr1GI
Professor Christopher Manning & Guest Speaker Kevin Clark, Stanford University
http://onlinehub.stanford.edu/
Professor Christopher Manning
Thomas M. Siebel Professor in Machine Learning, Professor of Linguistics and of Computer Science
Director, Stanford Artificial Intelligence Laboratory (SAIL)
To follow along with the course schedule and syllabus, visit: http://web.stanford.edu/class/....cs224n/index.html#sc
0:00 Introduction
0:56 Deep Learning for NLP 5 years ago
1:34 Future of Deep Learning + NLP
3:32 Why has deep learning been so successful recently?
5:03 Big deep learning successes
6:20 NLP Datasets
8:12 Machine Translation Data
9:54 Pre-Training
12:10 Self-Training
18:27 Large-Scale Back-Translation
20:03 Unsupervised Word Translation
27:43 Unsupervised Neural Machine Translation
31:14 Why Does This Work?
33:46 Unsupervised Machine Translation
34:49 Attribute Transfer
38:10 Cross-Lingual BERT
43:03 Huge Models in Computer Vision
44:29 Training Huge Models
47:41 So What Can GPT-2 Do?
50:04 GPT-2 Results
51:16 How can GPT-2 be doing translation?
52:12 GPT-2 Question Answering
53:31 What happens as models get even bigger?
54:24 GPT-2 Reaction
59:26 High-Impact Decisions