ComeçarComece de graça

Putting files in the cloud

Now that Sam knows how to create buckets, she is ready to automate a tedious part of her job. Right now, she has to download the latest files from the City of San Diego Open Data Portal, aggregate them, and share them with management.

Sharing an analysis with others is a common, yet tedious data science task. Automating these steps will allow Sam to focus on cooler projects, while keeping her management happy.

In the last lesson, Sam has already created the gid-staging bucket. She has already downloaded the files from the URLs, analyzed them, and wrote the results to final_report.csv.

She has also already initialized the boto3 S3 client and assigned it to the s3 variable.

Help Sam upload final_report.csv to the gid-staging bucket!

Este exercício faz parte do curso

Introduction to AWS Boto in Python

Ver curso

Instruções do exercício

  • Upload 'final_report.csv' to the 'gid-staging' bucket with the key '2019/final_report_01_01.csv'.
  • Get the object metadata and store it in response.
  • Print the object size in bytes.

Exercício interativo prático

Experimente este exercício completando este código de exemplo.

# Upload final_report.csv to gid-staging
s3.____(Bucket='____',
              # Set filename and key
               ____='____', 
               ____='____)

# Get object metadata and print it
response = s3.____(Bucket='gid-staging', 
                       Key='2019/final_report_01_01.csv')

# Print the size of the uploaded object
print(response['____'])
Editar e executar o código