Quiz 4 - Question 1
Assume you are training a small language model on the following 10 sentences that vary in length.
"I love NLP" (3 words)
"Natural language processing is fun" (5 words)
"Transformers are great" (3 words)
"We learn about n-grams and padding" (6 words)
"Probability plays a key role" (5 words)
"Batching helps with computational efficiency" (5 words)
"Padding sequences ensure uniform lengths" (5 words)
"Knowing your dataset is important" (5 words)
"Stochasticity introduces randomness" (3 words)
"Language models generate predictions" (4 words)
If you pad all sequences such that they are all as long as the longest sequence, how many pad tokens do you have to insert in total across all sequences?
Este exercício faz parte do curso
Google DeepMind: Build Your Own Small Language Model
Exercício interativo prático
Transforme a teoria em ação com um de nossos exercícios interativos
Começar o exercício