AdamW with Trainer
You're beginning to train a Transformer model to simplify language translations. As a first step, you decide to use the AdamW optimizer as a benchmark and the Trainer
interface for quick setup. Set up Trainer
to use the AdamW optimizer
.
AdamW
has been pre-imported from torch.optim
. Some training objects have been pre-loaded: model
, training_args
, train_dataset
, validation_dataset
, compute_metrics
.
This exercise is part of the course
Efficient AI Model Training with PyTorch
Exercise instructions
- Pass the
model
parameters to theAdamW
optimizer
. - Pass the
optimizer
toTrainer
.
Hands-on interactive exercise
Have a go at this exercise by completing this sample code.
# Pass the model parameters to the AdamW optimizer
optimizer = ____(params=____.____())
# Pass the optimizer to Trainer
trainer = Trainer(model=model,
args=training_args,
train_dataset=train_dataset,
eval_dataset=validation_dataset,
____=(____, None),
compute_metrics=compute_metrics)
trainer.train()