Get startedGet started for free

PyTorch Transformers

Now you're familiar with the different components of the transformer architecture, it's time to define one! The torch.nn module, imported for you as nn, provides a really nice way to do this in just a few lines of code.

This exercise is part of the course

Transformer Models with PyTorch

View Course

Exercise instructions

  • Define a transformer with 8 attention heads, 6 encoder and decoder layers, and for input sequence embeddings of length 1536.
  • Print the model object to view the model architecture.

Hands-on interactive exercise

Have a go at this exercise by completing this sample code.

# Define the transformer model
model = ____

# Print the model object
print(____)
Edit and Run Code