Quiz 4 - Question 1
Assume you are training a small language model on the following 10 sentences that vary in length.
"I love NLP" (3 words)
"Natural language processing is fun" (5 words)
"Transformers are great" (3 words)
"We learn about n-grams and padding" (6 words)
"Probability plays a key role" (5 words)
"Batching helps with computational efficiency" (5 words)
"Padding sequences ensure uniform lengths" (5 words)
"Knowing your dataset is important" (5 words)
"Stochasticity introduces randomness" (3 words)
"Language models generate predictions" (4 words)
If you pad all sequences such that they are all as long as the longest sequence, how many pad tokens do you have to insert in total across all sequences?
Cet exercice fait partie du cours
Google DeepMind: Build Your Own Small Language Model
Exercice interactif pratique
Passez de la théorie à la pratique avec l’un de nos exercices interactifs
Commencer l’exercice