Minsuk Heo 허민석
GPT-3 is super intelligent NLP deep learning model. In order to understand GPT-3 or later version, we should understand fundamental basic of it which is the GPT-2. I covered the how the GPT-2 could achieved zero shot learning and replaced high score on multiple NLP bench marks, and gave examples how GPT-2 as one model could achieve multiple NLP tasks without fine tuning.
Similar Posts
One thought on “GPT-2 (basic for understanding for GPT-3)”
Comments are closed.
Hello Sir,
if we take a queation similarity task, the input in Bert is:
CLS token + Question one + SEP + Question 2 + SEP
I read that yhe input in GPT-2 is:
Question one + Question 2 + CLS token.
Is this correct?
If yes,
Should we use the CLS token to represent the input for classification as we do in Bert?