Using the correct model structure
Question-answering can be either extractive or generative, each requiring a different transformer structure to process input and output correctly.
They use either:
- Encoder-only models such as
"distilbert-base-uncased-distilled-squad"
- Decoder-only models such as
"gpt2"
Use your knowledge of common models for specific tasks to select the appropriate one. pipeline is loaded, as well as text on the Mona Lisa.
This exercise is part of the course
Introduction to LLMs in Python
Hands-on interactive exercise
Have a go at this exercise by completing this sample code.
question = "Who painted the Mona Lisa?"
# Define the appropriate model
qa = pipeline(task="question-answering", model=____)
output = qa(question=question, context=text)
print(output['answer'])