Henry AI Labs
This video explains the original GPT model, “Improving Language Understanding by Generative Pre-Training”. I think the key takeaways are understanding that they use a new unlabeled text dataset that requires the pre-training language modeling to incorporate longer range context, the way that they format input representations for supervised fine-tuning, and the different NLP tasks this is evaluated on!
Paper Links:
GPT: https://s3-us-west-2.amazonaws.com/openai-assets/research-covers/language-unsupervised/language_understanding_paper.pdf
DeepMind “A new model and dataset for long range memory”: https://deepmind.com/blog/article/A_new_model_and_dataset_for_long-range_memory
SQuAD: https://rajpurkar.github.io/SQuAD-explorer/explore/v2.0/dev/Oxygen.html?model=BiDAF%20+%20Self%20Attention%20+%20ELMo%20(single%20model)%20(Allen%20Institute%20for%20Artificial%20Intelligence%20[modified%20by%20Stanford])&version=v2.0
MultiNLI: https://www.nyu.edu/projects/bowman/multinli/
RACE: https://arxiv.org/pdf/1704.04683.pdf
Quora Question Pairs: https://www.quora.com/q/quoradata/First-Quora-Dataset-Release-Question-Pairs
CoLA: https://arxiv.org/pdf/1805.12471.pdf
Thanks for watching! Please Subscribe!
0:55 Semi-Supervised Learning in NLP
1:25 BooksCorpus
2:30 Fine-Tuning Loss Function
3:30 Task-Specific Input Transformations
4:25 Transformer Decoder
4:45 Natural Language Inference
5:30 Question Answering
6:10 Semantic Similarity
6:40 CoLA Text Classification
6:48 All Tasks Tested
7:38 Ablations
Hey, thanks for your work. Can I ask what your tools and workflow look like for making such videos?
Congratulations on 10k subscribers.
I'd be happy if you'd do more videos on NLP a small wish from an early subscriber.
Good content!! Keep up the great work! Oh by the way, have a look at smzeus!!! Really think that it could help you grow your channel quicker.
Nice and Quick overview of the whole paper. Thanks.. 🙂