ComenzarEmpieza gratis

Quiz 4 - Question 1

Assume you are training a small language model on the following 10 sentences that vary in length.

"I love NLP" (3 words)
"Natural language processing is fun" (5 words)
"Transformers are great" (3 words)
"We learn about n-grams and padding" (6 words)
"Probability plays a key role" (5 words)
"Batching helps with computational efficiency" (5 words)
"Padding sequences ensure uniform lengths" (5 words)
"Knowing your dataset is important" (5 words)
"Stochasticity introduces randomness" (3 words)
"Language models generate predictions" (4 words)

If you pad all sequences such that they are all as long as the longest sequence, how many pad tokens do you have to insert in total across all sequences?

Este ejercicio forma parte del curso

Google DeepMind: Build Your Own Small Language Model

Ver curso

Ejercicio interactivo práctico

Pon en práctica la teoría con uno de nuestros ejercicios interactivos

Empezar ejercicio