2584 lines
89 KiB
Python
2584 lines
89 KiB
Python
# -*- coding: utf-8 -*-
|
|
from __future__ import unicode_literals
|
|
|
|
import datetime
|
|
from six.moves.urllib.request import urlopen
|
|
from six.moves.urllib.error import HTTPError
|
|
from functools import wraps
|
|
from gzip import GzipFile
|
|
from io import BytesIO
|
|
import zlib
|
|
|
|
import json
|
|
import boto
|
|
import boto3
|
|
from botocore.client import ClientError
|
|
import botocore.exceptions
|
|
from boto.exception import S3CreateError, S3ResponseError
|
|
from botocore.handlers import disable_signing
|
|
from boto.s3.connection import S3Connection
|
|
from boto.s3.key import Key
|
|
from freezegun import freeze_time
|
|
import six
|
|
import requests
|
|
import tests.backport_assert_raises # noqa
|
|
from nose.tools import assert_raises
|
|
|
|
import sure # noqa
|
|
|
|
from moto import settings, mock_s3, mock_s3_deprecated
|
|
import moto.s3.models as s3model
|
|
|
|
if settings.TEST_SERVER_MODE:
|
|
REDUCED_PART_SIZE = s3model.UPLOAD_PART_MIN_SIZE
|
|
EXPECTED_ETAG = '"140f92a6df9f9e415f74a1463bcee9bb-2"'
|
|
else:
|
|
REDUCED_PART_SIZE = 256
|
|
EXPECTED_ETAG = '"66d1a1a2ed08fd05c137f316af4ff255-2"'
|
|
|
|
|
|
def reduced_min_part_size(f):
|
|
""" speed up tests by temporarily making the multipart minimum part size
|
|
small
|
|
"""
|
|
orig_size = s3model.UPLOAD_PART_MIN_SIZE
|
|
|
|
@wraps(f)
|
|
def wrapped(*args, **kwargs):
|
|
try:
|
|
s3model.UPLOAD_PART_MIN_SIZE = REDUCED_PART_SIZE
|
|
return f(*args, **kwargs)
|
|
finally:
|
|
s3model.UPLOAD_PART_MIN_SIZE = orig_size
|
|
|
|
return wrapped
|
|
|
|
|
|
class MyModel(object):
|
|
|
|
def __init__(self, name, value):
|
|
self.name = name
|
|
self.value = value
|
|
|
|
def save(self):
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
s3.put_object(Bucket='mybucket', Key=self.name, Body=self.value)
|
|
|
|
|
|
@mock_s3
|
|
def test_my_model_save():
|
|
# Create Bucket so that test can run
|
|
conn = boto3.resource('s3', region_name='us-east-1')
|
|
conn.create_bucket(Bucket='mybucket')
|
|
####################################
|
|
|
|
model_instance = MyModel('steve', 'is awesome')
|
|
model_instance.save()
|
|
|
|
body = conn.Object('mybucket', 'steve').get()['Body'].read().decode()
|
|
|
|
assert body == 'is awesome'
|
|
|
|
|
|
@mock_s3
|
|
def test_key_etag():
|
|
conn = boto3.resource('s3', region_name='us-east-1')
|
|
conn.create_bucket(Bucket='mybucket')
|
|
|
|
model_instance = MyModel('steve', 'is awesome')
|
|
model_instance.save()
|
|
|
|
conn.Bucket('mybucket').Object('steve').e_tag.should.equal(
|
|
'"d32bda93738f7e03adb22e66c90fbc04"')
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_multipart_upload_too_small():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
|
|
multipart = bucket.initiate_multipart_upload("the-key")
|
|
multipart.upload_part_from_file(BytesIO(b'hello'), 1)
|
|
multipart.upload_part_from_file(BytesIO(b'world'), 2)
|
|
# Multipart with total size under 5MB is refused
|
|
multipart.complete_upload.should.throw(S3ResponseError)
|
|
|
|
|
|
@mock_s3_deprecated
|
|
@reduced_min_part_size
|
|
def test_multipart_upload():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
|
|
multipart = bucket.initiate_multipart_upload("the-key")
|
|
part1 = b'0' * REDUCED_PART_SIZE
|
|
multipart.upload_part_from_file(BytesIO(part1), 1)
|
|
# last part, can be less than 5 MB
|
|
part2 = b'1'
|
|
multipart.upload_part_from_file(BytesIO(part2), 2)
|
|
multipart.complete_upload()
|
|
# we should get both parts as the key contents
|
|
bucket.get_key(
|
|
"the-key").get_contents_as_string().should.equal(part1 + part2)
|
|
|
|
|
|
@mock_s3_deprecated
|
|
@reduced_min_part_size
|
|
def test_multipart_upload_out_of_order():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
|
|
multipart = bucket.initiate_multipart_upload("the-key")
|
|
# last part, can be less than 5 MB
|
|
part2 = b'1'
|
|
multipart.upload_part_from_file(BytesIO(part2), 4)
|
|
part1 = b'0' * REDUCED_PART_SIZE
|
|
multipart.upload_part_from_file(BytesIO(part1), 2)
|
|
multipart.complete_upload()
|
|
# we should get both parts as the key contents
|
|
bucket.get_key(
|
|
"the-key").get_contents_as_string().should.equal(part1 + part2)
|
|
|
|
|
|
@mock_s3_deprecated
|
|
@reduced_min_part_size
|
|
def test_multipart_upload_with_headers():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
|
|
multipart = bucket.initiate_multipart_upload(
|
|
"the-key", metadata={"foo": "bar"})
|
|
part1 = b'0' * 10
|
|
multipart.upload_part_from_file(BytesIO(part1), 1)
|
|
multipart.complete_upload()
|
|
|
|
key = bucket.get_key("the-key")
|
|
key.metadata.should.equal({"foo": "bar"})
|
|
|
|
|
|
@mock_s3_deprecated
|
|
@reduced_min_part_size
|
|
def test_multipart_upload_with_copy_key():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
key = Key(bucket)
|
|
key.key = "original-key"
|
|
key.set_contents_from_string("key_value")
|
|
|
|
multipart = bucket.initiate_multipart_upload("the-key")
|
|
part1 = b'0' * REDUCED_PART_SIZE
|
|
multipart.upload_part_from_file(BytesIO(part1), 1)
|
|
multipart.copy_part_from_key("foobar", "original-key", 2, 0, 3)
|
|
multipart.complete_upload()
|
|
bucket.get_key(
|
|
"the-key").get_contents_as_string().should.equal(part1 + b"key_")
|
|
|
|
|
|
@mock_s3_deprecated
|
|
@reduced_min_part_size
|
|
def test_multipart_upload_cancel():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
|
|
multipart = bucket.initiate_multipart_upload("the-key")
|
|
part1 = b'0' * REDUCED_PART_SIZE
|
|
multipart.upload_part_from_file(BytesIO(part1), 1)
|
|
multipart.cancel_upload()
|
|
# TODO we really need some sort of assertion here, but we don't currently
|
|
# have the ability to list mulipart uploads for a bucket.
|
|
|
|
|
|
@mock_s3_deprecated
|
|
@reduced_min_part_size
|
|
def test_multipart_etag():
|
|
# Create Bucket so that test can run
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket('mybucket')
|
|
|
|
multipart = bucket.initiate_multipart_upload("the-key")
|
|
part1 = b'0' * REDUCED_PART_SIZE
|
|
multipart.upload_part_from_file(BytesIO(part1), 1)
|
|
# last part, can be less than 5 MB
|
|
part2 = b'1'
|
|
multipart.upload_part_from_file(BytesIO(part2), 2)
|
|
multipart.complete_upload()
|
|
# we should get both parts as the key contents
|
|
bucket.get_key("the-key").etag.should.equal(EXPECTED_ETAG)
|
|
|
|
|
|
@mock_s3_deprecated
|
|
@reduced_min_part_size
|
|
def test_multipart_invalid_order():
|
|
# Create Bucket so that test can run
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket('mybucket')
|
|
|
|
multipart = bucket.initiate_multipart_upload("the-key")
|
|
part1 = b'0' * 5242880
|
|
etag1 = multipart.upload_part_from_file(BytesIO(part1), 1).etag
|
|
# last part, can be less than 5 MB
|
|
part2 = b'1'
|
|
etag2 = multipart.upload_part_from_file(BytesIO(part2), 2).etag
|
|
xml = "<Part><PartNumber>{0}</PartNumber><ETag>{1}</ETag></Part>"
|
|
xml = xml.format(2, etag2) + xml.format(1, etag1)
|
|
xml = "<CompleteMultipartUpload>{0}</CompleteMultipartUpload>".format(xml)
|
|
bucket.complete_multipart_upload.when.called_with(
|
|
multipart.key_name, multipart.id, xml).should.throw(S3ResponseError)
|
|
|
|
@mock_s3_deprecated
|
|
@reduced_min_part_size
|
|
def test_multipart_etag_quotes_stripped():
|
|
# Create Bucket so that test can run
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket('mybucket')
|
|
|
|
multipart = bucket.initiate_multipart_upload("the-key")
|
|
part1 = b'0' * REDUCED_PART_SIZE
|
|
etag1 = multipart.upload_part_from_file(BytesIO(part1), 1).etag
|
|
# last part, can be less than 5 MB
|
|
part2 = b'1'
|
|
etag2 = multipart.upload_part_from_file(BytesIO(part2), 2).etag
|
|
# Strip quotes from etags
|
|
etag1 = etag1.replace('"','')
|
|
etag2 = etag2.replace('"','')
|
|
xml = "<Part><PartNumber>{0}</PartNumber><ETag>{1}</ETag></Part>"
|
|
xml = xml.format(1, etag1) + xml.format(2, etag2)
|
|
xml = "<CompleteMultipartUpload>{0}</CompleteMultipartUpload>".format(xml)
|
|
bucket.complete_multipart_upload.when.called_with(
|
|
multipart.key_name, multipart.id, xml).should_not.throw(S3ResponseError)
|
|
# we should get both parts as the key contents
|
|
bucket.get_key("the-key").etag.should.equal(EXPECTED_ETAG)
|
|
|
|
@mock_s3_deprecated
|
|
@reduced_min_part_size
|
|
def test_multipart_duplicate_upload():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
|
|
multipart = bucket.initiate_multipart_upload("the-key")
|
|
part1 = b'0' * REDUCED_PART_SIZE
|
|
multipart.upload_part_from_file(BytesIO(part1), 1)
|
|
# same part again
|
|
multipart.upload_part_from_file(BytesIO(part1), 1)
|
|
part2 = b'1' * 1024
|
|
multipart.upload_part_from_file(BytesIO(part2), 2)
|
|
multipart.complete_upload()
|
|
# We should get only one copy of part 1.
|
|
bucket.get_key(
|
|
"the-key").get_contents_as_string().should.equal(part1 + part2)
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_list_multiparts():
|
|
# Create Bucket so that test can run
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket('mybucket')
|
|
|
|
multipart1 = bucket.initiate_multipart_upload("one-key")
|
|
multipart2 = bucket.initiate_multipart_upload("two-key")
|
|
uploads = bucket.get_all_multipart_uploads()
|
|
uploads.should.have.length_of(2)
|
|
dict([(u.key_name, u.id) for u in uploads]).should.equal(
|
|
{'one-key': multipart1.id, 'two-key': multipart2.id})
|
|
multipart2.cancel_upload()
|
|
uploads = bucket.get_all_multipart_uploads()
|
|
uploads.should.have.length_of(1)
|
|
uploads[0].key_name.should.equal("one-key")
|
|
multipart1.cancel_upload()
|
|
uploads = bucket.get_all_multipart_uploads()
|
|
uploads.should.be.empty
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_key_save_to_missing_bucket():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.get_bucket('mybucket', validate=False)
|
|
|
|
key = Key(bucket)
|
|
key.key = "the-key"
|
|
key.set_contents_from_string.when.called_with(
|
|
"foobar").should.throw(S3ResponseError)
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_missing_key():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
bucket.get_key("the-key").should.equal(None)
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_missing_key_urllib2():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
conn.create_bucket("foobar")
|
|
|
|
urlopen.when.called_with(
|
|
"http://foobar.s3.amazonaws.com/the-key").should.throw(HTTPError)
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_empty_key():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
key = Key(bucket)
|
|
key.key = "the-key"
|
|
key.set_contents_from_string("")
|
|
|
|
key = bucket.get_key("the-key")
|
|
key.size.should.equal(0)
|
|
key.get_contents_as_string().should.equal(b'')
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_empty_key_set_on_existing_key():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
key = Key(bucket)
|
|
key.key = "the-key"
|
|
key.set_contents_from_string("foobar")
|
|
|
|
key = bucket.get_key("the-key")
|
|
key.size.should.equal(6)
|
|
key.get_contents_as_string().should.equal(b'foobar')
|
|
|
|
key.set_contents_from_string("")
|
|
bucket.get_key("the-key").get_contents_as_string().should.equal(b'')
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_large_key_save():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
key = Key(bucket)
|
|
key.key = "the-key"
|
|
key.set_contents_from_string("foobar" * 100000)
|
|
|
|
bucket.get_key(
|
|
"the-key").get_contents_as_string().should.equal(b'foobar' * 100000)
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_copy_key():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
key = Key(bucket)
|
|
key.key = "the-key"
|
|
key.set_contents_from_string("some value")
|
|
|
|
bucket.copy_key('new-key', 'foobar', 'the-key')
|
|
|
|
bucket.get_key(
|
|
"the-key").get_contents_as_string().should.equal(b"some value")
|
|
bucket.get_key(
|
|
"new-key").get_contents_as_string().should.equal(b"some value")
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_copy_key_with_version():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
bucket.configure_versioning(versioning=True)
|
|
key = Key(bucket)
|
|
key.key = "the-key"
|
|
key.set_contents_from_string("some value")
|
|
key.set_contents_from_string("another value")
|
|
|
|
bucket.copy_key('new-key', 'foobar', 'the-key', src_version_id='0')
|
|
|
|
bucket.get_key(
|
|
"the-key").get_contents_as_string().should.equal(b"another value")
|
|
bucket.get_key(
|
|
"new-key").get_contents_as_string().should.equal(b"some value")
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_set_metadata():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
key = Key(bucket)
|
|
key.key = 'the-key'
|
|
key.set_metadata('md', 'Metadatastring')
|
|
key.set_contents_from_string("Testval")
|
|
|
|
bucket.get_key('the-key').get_metadata('md').should.equal('Metadatastring')
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_copy_key_replace_metadata():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
key = Key(bucket)
|
|
key.key = "the-key"
|
|
key.set_metadata('md', 'Metadatastring')
|
|
key.set_contents_from_string("some value")
|
|
|
|
bucket.copy_key('new-key', 'foobar', 'the-key',
|
|
metadata={'momd': 'Mometadatastring'})
|
|
|
|
bucket.get_key("new-key").get_metadata('md').should.be.none
|
|
bucket.get_key(
|
|
"new-key").get_metadata('momd').should.equal('Mometadatastring')
|
|
|
|
|
|
@freeze_time("2012-01-01 12:00:00")
|
|
@mock_s3_deprecated
|
|
def test_last_modified():
|
|
# See https://github.com/boto/boto/issues/466
|
|
conn = boto.connect_s3()
|
|
bucket = conn.create_bucket("foobar")
|
|
key = Key(bucket)
|
|
key.key = "the-key"
|
|
key.set_contents_from_string("some value")
|
|
|
|
rs = bucket.get_all_keys()
|
|
rs[0].last_modified.should.equal('2012-01-01T12:00:00.000Z')
|
|
|
|
bucket.get_key(
|
|
"the-key").last_modified.should.equal('Sun, 01 Jan 2012 12:00:00 GMT')
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_missing_bucket():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
conn.get_bucket.when.called_with('mybucket').should.throw(S3ResponseError)
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_bucket_with_dash():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
conn.get_bucket.when.called_with(
|
|
'mybucket-test').should.throw(S3ResponseError)
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_create_existing_bucket():
|
|
"Trying to create a bucket that already exists should raise an Error"
|
|
conn = boto.s3.connect_to_region("us-west-2")
|
|
conn.create_bucket("foobar")
|
|
with assert_raises(S3CreateError):
|
|
conn.create_bucket('foobar')
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_create_existing_bucket_in_us_east_1():
|
|
"Trying to create a bucket that already exists in us-east-1 returns the bucket"
|
|
|
|
""""
|
|
http://docs.aws.amazon.com/AmazonS3/latest/API/ErrorResponses.html
|
|
Your previous request to create the named bucket succeeded and you already
|
|
own it. You get this error in all AWS regions except US Standard,
|
|
us-east-1. In us-east-1 region, you will get 200 OK, but it is no-op (if
|
|
bucket exists it Amazon S3 will not do anything).
|
|
"""
|
|
conn = boto.s3.connect_to_region("us-east-1")
|
|
conn.create_bucket("foobar")
|
|
bucket = conn.create_bucket("foobar")
|
|
bucket.name.should.equal("foobar")
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_other_region():
|
|
conn = S3Connection(
|
|
'key', 'secret', host='s3-website-ap-southeast-2.amazonaws.com')
|
|
conn.create_bucket("foobar")
|
|
list(conn.get_bucket("foobar").get_all_keys()).should.equal([])
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_bucket_deletion():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
|
|
key = Key(bucket)
|
|
key.key = "the-key"
|
|
key.set_contents_from_string("some value")
|
|
|
|
# Try to delete a bucket that still has keys
|
|
conn.delete_bucket.when.called_with("foobar").should.throw(S3ResponseError)
|
|
|
|
bucket.delete_key("the-key")
|
|
conn.delete_bucket("foobar")
|
|
|
|
# Get non-existing bucket
|
|
conn.get_bucket.when.called_with("foobar").should.throw(S3ResponseError)
|
|
|
|
# Delete non-existant bucket
|
|
conn.delete_bucket.when.called_with("foobar").should.throw(S3ResponseError)
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_get_all_buckets():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
conn.create_bucket("foobar")
|
|
conn.create_bucket("foobar2")
|
|
buckets = conn.get_all_buckets()
|
|
|
|
buckets.should.have.length_of(2)
|
|
|
|
|
|
@mock_s3
|
|
@mock_s3_deprecated
|
|
def test_post_to_bucket():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
|
|
requests.post("https://foobar.s3.amazonaws.com/", {
|
|
'key': 'the-key',
|
|
'file': 'nothing'
|
|
})
|
|
|
|
bucket.get_key('the-key').get_contents_as_string().should.equal(b'nothing')
|
|
|
|
|
|
@mock_s3
|
|
@mock_s3_deprecated
|
|
def test_post_with_metadata_to_bucket():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
|
|
requests.post("https://foobar.s3.amazonaws.com/", {
|
|
'key': 'the-key',
|
|
'file': 'nothing',
|
|
'x-amz-meta-test': 'metadata'
|
|
})
|
|
|
|
bucket.get_key('the-key').get_metadata('test').should.equal('metadata')
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_delete_missing_key():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket('foobar')
|
|
|
|
deleted_key = bucket.delete_key("foobar")
|
|
deleted_key.key.should.equal("foobar")
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_delete_keys():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket('foobar')
|
|
|
|
Key(bucket=bucket, name='file1').set_contents_from_string('abc')
|
|
Key(bucket=bucket, name='file2').set_contents_from_string('abc')
|
|
Key(bucket=bucket, name='file3').set_contents_from_string('abc')
|
|
Key(bucket=bucket, name='file4').set_contents_from_string('abc')
|
|
|
|
result = bucket.delete_keys(['file2', 'file3'])
|
|
result.deleted.should.have.length_of(2)
|
|
result.errors.should.have.length_of(0)
|
|
keys = bucket.get_all_keys()
|
|
keys.should.have.length_of(2)
|
|
keys[0].name.should.equal('file1')
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_delete_keys_with_invalid():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket('foobar')
|
|
|
|
Key(bucket=bucket, name='file1').set_contents_from_string('abc')
|
|
Key(bucket=bucket, name='file2').set_contents_from_string('abc')
|
|
Key(bucket=bucket, name='file3').set_contents_from_string('abc')
|
|
Key(bucket=bucket, name='file4').set_contents_from_string('abc')
|
|
|
|
result = bucket.delete_keys(['abc', 'file3'])
|
|
|
|
result.deleted.should.have.length_of(1)
|
|
result.errors.should.have.length_of(1)
|
|
keys = bucket.get_all_keys()
|
|
keys.should.have.length_of(3)
|
|
keys[0].name.should.equal('file1')
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_bucket_name_with_dot():
|
|
conn = boto.connect_s3()
|
|
bucket = conn.create_bucket('firstname.lastname')
|
|
|
|
k = Key(bucket, 'somekey')
|
|
k.set_contents_from_string('somedata')
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_key_with_special_characters():
|
|
conn = boto.connect_s3()
|
|
bucket = conn.create_bucket('test_bucket_name')
|
|
|
|
key = Key(bucket, 'test_list_keys_2/x?y')
|
|
key.set_contents_from_string('value1')
|
|
|
|
key_list = bucket.list('test_list_keys_2/', '/')
|
|
keys = [x for x in key_list]
|
|
keys[0].name.should.equal("test_list_keys_2/x?y")
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_unicode_key_with_slash():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
key = Key(bucket)
|
|
key.key = "/the-key-unîcode/test"
|
|
key.set_contents_from_string("value")
|
|
|
|
key = bucket.get_key("/the-key-unîcode/test")
|
|
key.get_contents_as_string().should.equal(b'value')
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_bucket_key_listing_order():
|
|
conn = boto.connect_s3()
|
|
bucket = conn.create_bucket('test_bucket')
|
|
prefix = 'toplevel/'
|
|
|
|
def store(name):
|
|
k = Key(bucket, prefix + name)
|
|
k.set_contents_from_string('somedata')
|
|
|
|
names = ['x/key', 'y.key1', 'y.key2', 'y.key3', 'x/y/key', 'x/y/z/key']
|
|
|
|
for name in names:
|
|
store(name)
|
|
|
|
delimiter = None
|
|
keys = [x.name for x in bucket.list(prefix, delimiter)]
|
|
keys.should.equal([
|
|
'toplevel/x/key', 'toplevel/x/y/key', 'toplevel/x/y/z/key',
|
|
'toplevel/y.key1', 'toplevel/y.key2', 'toplevel/y.key3'
|
|
])
|
|
|
|
delimiter = '/'
|
|
keys = [x.name for x in bucket.list(prefix, delimiter)]
|
|
keys.should.equal([
|
|
'toplevel/y.key1', 'toplevel/y.key2', 'toplevel/y.key3', 'toplevel/x/'
|
|
])
|
|
|
|
# Test delimiter with no prefix
|
|
delimiter = '/'
|
|
keys = [x.name for x in bucket.list(prefix=None, delimiter=delimiter)]
|
|
keys.should.equal(['toplevel/'])
|
|
|
|
delimiter = None
|
|
keys = [x.name for x in bucket.list(prefix + 'x', delimiter)]
|
|
keys.should.equal(
|
|
[u'toplevel/x/key', u'toplevel/x/y/key', u'toplevel/x/y/z/key'])
|
|
|
|
delimiter = '/'
|
|
keys = [x.name for x in bucket.list(prefix + 'x', delimiter)]
|
|
keys.should.equal([u'toplevel/x/'])
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_key_with_reduced_redundancy():
|
|
conn = boto.connect_s3()
|
|
bucket = conn.create_bucket('test_bucket_name')
|
|
|
|
key = Key(bucket, 'test_rr_key')
|
|
key.set_contents_from_string('value1', reduced_redundancy=True)
|
|
# we use the bucket iterator because of:
|
|
# https:/github.com/boto/boto/issues/1173
|
|
list(bucket)[0].storage_class.should.equal('REDUCED_REDUNDANCY')
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_copy_key_reduced_redundancy():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
key = Key(bucket)
|
|
key.key = "the-key"
|
|
key.set_contents_from_string("some value")
|
|
|
|
bucket.copy_key('new-key', 'foobar', 'the-key',
|
|
storage_class='REDUCED_REDUNDANCY')
|
|
|
|
# we use the bucket iterator because of:
|
|
# https:/github.com/boto/boto/issues/1173
|
|
keys = dict([(k.name, k) for k in bucket])
|
|
keys['new-key'].storage_class.should.equal("REDUCED_REDUNDANCY")
|
|
keys['the-key'].storage_class.should.equal("STANDARD")
|
|
|
|
|
|
@freeze_time("2012-01-01 12:00:00")
|
|
@mock_s3_deprecated
|
|
def test_restore_key():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
key = Key(bucket)
|
|
key.key = "the-key"
|
|
key.set_contents_from_string("some value")
|
|
list(bucket)[0].ongoing_restore.should.be.none
|
|
key.restore(1)
|
|
key = bucket.get_key('the-key')
|
|
key.ongoing_restore.should_not.be.none
|
|
key.ongoing_restore.should.be.false
|
|
key.expiry_date.should.equal("Mon, 02 Jan 2012 12:00:00 GMT")
|
|
key.restore(2)
|
|
key = bucket.get_key('the-key')
|
|
key.ongoing_restore.should_not.be.none
|
|
key.ongoing_restore.should.be.false
|
|
key.expiry_date.should.equal("Tue, 03 Jan 2012 12:00:00 GMT")
|
|
|
|
|
|
@freeze_time("2012-01-01 12:00:00")
|
|
@mock_s3_deprecated
|
|
def test_restore_key_headers():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket("foobar")
|
|
key = Key(bucket)
|
|
key.key = "the-key"
|
|
key.set_contents_from_string("some value")
|
|
key.restore(1, headers={'foo': 'bar'})
|
|
key = bucket.get_key('the-key')
|
|
key.ongoing_restore.should_not.be.none
|
|
key.ongoing_restore.should.be.false
|
|
key.expiry_date.should.equal("Mon, 02 Jan 2012 12:00:00 GMT")
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_get_versioning_status():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket('foobar')
|
|
d = bucket.get_versioning_status()
|
|
d.should.be.empty
|
|
|
|
bucket.configure_versioning(versioning=True)
|
|
d = bucket.get_versioning_status()
|
|
d.shouldnt.be.empty
|
|
d.should.have.key('Versioning').being.equal('Enabled')
|
|
|
|
bucket.configure_versioning(versioning=False)
|
|
d = bucket.get_versioning_status()
|
|
d.should.have.key('Versioning').being.equal('Suspended')
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_key_version():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket('foobar')
|
|
bucket.configure_versioning(versioning=True)
|
|
|
|
key = Key(bucket)
|
|
key.key = 'the-key'
|
|
key.version_id.should.be.none
|
|
key.set_contents_from_string('some string')
|
|
key.version_id.should.equal('0')
|
|
key.set_contents_from_string('some string')
|
|
key.version_id.should.equal('1')
|
|
|
|
key = bucket.get_key('the-key')
|
|
key.version_id.should.equal('1')
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_list_versions():
|
|
conn = boto.connect_s3('the_key', 'the_secret')
|
|
bucket = conn.create_bucket('foobar')
|
|
bucket.configure_versioning(versioning=True)
|
|
|
|
key = Key(bucket, 'the-key')
|
|
key.version_id.should.be.none
|
|
key.set_contents_from_string("Version 1")
|
|
key.version_id.should.equal('0')
|
|
key.set_contents_from_string("Version 2")
|
|
key.version_id.should.equal('1')
|
|
|
|
versions = list(bucket.list_versions())
|
|
|
|
versions.should.have.length_of(2)
|
|
|
|
versions[0].name.should.equal('the-key')
|
|
versions[0].version_id.should.equal('0')
|
|
versions[0].get_contents_as_string().should.equal(b"Version 1")
|
|
|
|
versions[1].name.should.equal('the-key')
|
|
versions[1].version_id.should.equal('1')
|
|
versions[1].get_contents_as_string().should.equal(b"Version 2")
|
|
|
|
key = Key(bucket, 'the2-key')
|
|
key.set_contents_from_string("Version 1")
|
|
|
|
keys = list(bucket.list())
|
|
keys.should.have.length_of(2)
|
|
versions = list(bucket.list_versions(prefix='the2-'))
|
|
versions.should.have.length_of(1)
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_acl_setting():
|
|
conn = boto.connect_s3()
|
|
bucket = conn.create_bucket('foobar')
|
|
content = b'imafile'
|
|
keyname = 'test.txt'
|
|
|
|
key = Key(bucket, name=keyname)
|
|
key.content_type = 'text/plain'
|
|
key.set_contents_from_string(content)
|
|
key.make_public()
|
|
|
|
key = bucket.get_key(keyname)
|
|
|
|
assert key.get_contents_as_string() == content
|
|
|
|
grants = key.get_acl().acl.grants
|
|
assert any(g.uri == 'http://acs.amazonaws.com/groups/global/AllUsers' and
|
|
g.permission == 'READ' for g in grants), grants
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_acl_setting_via_headers():
|
|
conn = boto.connect_s3()
|
|
bucket = conn.create_bucket('foobar')
|
|
content = b'imafile'
|
|
keyname = 'test.txt'
|
|
|
|
key = Key(bucket, name=keyname)
|
|
key.content_type = 'text/plain'
|
|
key.set_contents_from_string(content, headers={
|
|
'x-amz-grant-full-control': 'uri="http://acs.amazonaws.com/groups/global/AllUsers"'
|
|
})
|
|
|
|
key = bucket.get_key(keyname)
|
|
|
|
assert key.get_contents_as_string() == content
|
|
|
|
grants = key.get_acl().acl.grants
|
|
assert any(g.uri == 'http://acs.amazonaws.com/groups/global/AllUsers' and
|
|
g.permission == 'FULL_CONTROL' for g in grants), grants
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_acl_switching():
|
|
conn = boto.connect_s3()
|
|
bucket = conn.create_bucket('foobar')
|
|
content = b'imafile'
|
|
keyname = 'test.txt'
|
|
|
|
key = Key(bucket, name=keyname)
|
|
key.content_type = 'text/plain'
|
|
key.set_contents_from_string(content, policy='public-read')
|
|
key.set_acl('private')
|
|
|
|
grants = key.get_acl().acl.grants
|
|
assert not any(g.uri == 'http://acs.amazonaws.com/groups/global/AllUsers' and
|
|
g.permission == 'READ' for g in grants), grants
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_bucket_acl_setting():
|
|
conn = boto.connect_s3()
|
|
bucket = conn.create_bucket('foobar')
|
|
|
|
bucket.make_public()
|
|
|
|
grants = bucket.get_acl().acl.grants
|
|
assert any(g.uri == 'http://acs.amazonaws.com/groups/global/AllUsers' and
|
|
g.permission == 'READ' for g in grants), grants
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_bucket_acl_switching():
|
|
conn = boto.connect_s3()
|
|
bucket = conn.create_bucket('foobar')
|
|
bucket.make_public()
|
|
|
|
bucket.set_acl('private')
|
|
|
|
grants = bucket.get_acl().acl.grants
|
|
assert not any(g.uri == 'http://acs.amazonaws.com/groups/global/AllUsers' and
|
|
g.permission == 'READ' for g in grants), grants
|
|
|
|
|
|
@mock_s3
|
|
def test_s3_object_in_public_bucket():
|
|
s3 = boto3.resource('s3')
|
|
bucket = s3.Bucket('test-bucket')
|
|
bucket.create(ACL='public-read')
|
|
bucket.put_object(Body=b'ABCD', Key='file.txt')
|
|
|
|
s3_anonymous = boto3.resource('s3')
|
|
s3_anonymous.meta.client.meta.events.register('choose-signer.s3.*', disable_signing)
|
|
|
|
contents = s3_anonymous.Object(key='file.txt', bucket_name='test-bucket').get()['Body'].read()
|
|
contents.should.equal(b'ABCD')
|
|
|
|
bucket.put_object(ACL='private', Body=b'ABCD', Key='file.txt')
|
|
|
|
with assert_raises(ClientError) as exc:
|
|
s3_anonymous.Object(key='file.txt', bucket_name='test-bucket').get()
|
|
exc.exception.response['Error']['Code'].should.equal('403')
|
|
|
|
params = {'Bucket': 'test-bucket', 'Key': 'file.txt'}
|
|
presigned_url = boto3.client('s3').generate_presigned_url('get_object', params, ExpiresIn=900)
|
|
response = requests.get(presigned_url)
|
|
assert response.status_code == 200
|
|
|
|
|
|
@mock_s3
|
|
def test_s3_object_in_private_bucket():
|
|
s3 = boto3.resource('s3')
|
|
bucket = s3.Bucket('test-bucket')
|
|
bucket.create(ACL='private')
|
|
bucket.put_object(ACL='private', Body=b'ABCD', Key='file.txt')
|
|
|
|
s3_anonymous = boto3.resource('s3')
|
|
s3_anonymous.meta.client.meta.events.register('choose-signer.s3.*', disable_signing)
|
|
|
|
with assert_raises(ClientError) as exc:
|
|
s3_anonymous.Object(key='file.txt', bucket_name='test-bucket').get()
|
|
exc.exception.response['Error']['Code'].should.equal('403')
|
|
|
|
bucket.put_object(ACL='public-read', Body=b'ABCD', Key='file.txt')
|
|
contents = s3_anonymous.Object(key='file.txt', bucket_name='test-bucket').get()['Body'].read()
|
|
contents.should.equal(b'ABCD')
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_unicode_key():
|
|
conn = boto.connect_s3()
|
|
bucket = conn.create_bucket('mybucket')
|
|
key = Key(bucket)
|
|
key.key = u'こんにちは.jpg'
|
|
key.set_contents_from_string('Hello world!')
|
|
assert [listed_key.key for listed_key in bucket.list()] == [key.key]
|
|
fetched_key = bucket.get_key(key.key)
|
|
assert fetched_key.key == key.key
|
|
assert fetched_key.get_contents_as_string().decode("utf-8") == 'Hello world!'
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_unicode_value():
|
|
conn = boto.connect_s3()
|
|
bucket = conn.create_bucket('mybucket')
|
|
key = Key(bucket)
|
|
key.key = 'some_key'
|
|
key.set_contents_from_string(u'こんにちは.jpg')
|
|
list(bucket.list())
|
|
key = bucket.get_key(key.key)
|
|
assert key.get_contents_as_string().decode("utf-8") == u'こんにちは.jpg'
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_setting_content_encoding():
|
|
conn = boto.connect_s3()
|
|
bucket = conn.create_bucket('mybucket')
|
|
key = bucket.new_key("keyname")
|
|
key.set_metadata("Content-Encoding", "gzip")
|
|
compressed_data = "abcdef"
|
|
key.set_contents_from_string(compressed_data)
|
|
|
|
key = bucket.get_key("keyname")
|
|
key.content_encoding.should.equal("gzip")
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_bucket_location():
|
|
conn = boto.s3.connect_to_region("us-west-2")
|
|
bucket = conn.create_bucket('mybucket')
|
|
bucket.get_location().should.equal("us-west-2")
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_ranged_get():
|
|
conn = boto.connect_s3()
|
|
bucket = conn.create_bucket('mybucket')
|
|
key = Key(bucket)
|
|
key.key = 'bigkey'
|
|
rep = b"0123456789"
|
|
key.set_contents_from_string(rep * 10)
|
|
|
|
# Implicitly bounded range requests.
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=0-'}).should.equal(rep * 10)
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=50-'}).should.equal(rep * 5)
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=99-'}).should.equal(b'9')
|
|
|
|
# Explicitly bounded range requests starting from the first byte.
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=0-0'}).should.equal(b'0')
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=0-49'}).should.equal(rep * 5)
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=0-99'}).should.equal(rep * 10)
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=0-100'}).should.equal(rep * 10)
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=0-700'}).should.equal(rep * 10)
|
|
|
|
# Explicitly bounded range requests starting from the / a middle byte.
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=50-54'}).should.equal(rep[:5])
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=50-99'}).should.equal(rep * 5)
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=50-100'}).should.equal(rep * 5)
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=50-700'}).should.equal(rep * 5)
|
|
|
|
# Explicitly bounded range requests starting from the last byte.
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=99-99'}).should.equal(b'9')
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=99-100'}).should.equal(b'9')
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=99-700'}).should.equal(b'9')
|
|
|
|
# Suffix range requests.
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=-1'}).should.equal(b'9')
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=-60'}).should.equal(rep * 6)
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=-100'}).should.equal(rep * 10)
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=-101'}).should.equal(rep * 10)
|
|
key.get_contents_as_string(
|
|
headers={'Range': 'bytes=-700'}).should.equal(rep * 10)
|
|
|
|
key.size.should.equal(100)
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_policy():
|
|
conn = boto.connect_s3()
|
|
bucket_name = 'mybucket'
|
|
bucket = conn.create_bucket(bucket_name)
|
|
|
|
policy = json.dumps({
|
|
"Version": "2012-10-17",
|
|
"Id": "PutObjPolicy",
|
|
"Statement": [
|
|
{
|
|
"Sid": "DenyUnEncryptedObjectUploads",
|
|
"Effect": "Deny",
|
|
"Principal": "*",
|
|
"Action": "s3:PutObject",
|
|
"Resource": "arn:aws:s3:::{bucket_name}/*".format(bucket_name=bucket_name),
|
|
"Condition": {
|
|
"StringNotEquals": {
|
|
"s3:x-amz-server-side-encryption": "aws:kms"
|
|
}
|
|
}
|
|
}
|
|
]
|
|
})
|
|
|
|
with assert_raises(S3ResponseError) as err:
|
|
bucket.get_policy()
|
|
|
|
ex = err.exception
|
|
ex.box_usage.should.be.none
|
|
ex.error_code.should.equal('NoSuchBucketPolicy')
|
|
ex.message.should.equal('The bucket policy does not exist')
|
|
ex.reason.should.equal('Not Found')
|
|
ex.resource.should.be.none
|
|
ex.status.should.equal(404)
|
|
ex.body.should.contain(bucket_name)
|
|
ex.request_id.should_not.be.none
|
|
|
|
bucket.set_policy(policy).should.be.true
|
|
|
|
bucket = conn.get_bucket(bucket_name)
|
|
|
|
bucket.get_policy().decode('utf-8').should.equal(policy)
|
|
|
|
bucket.delete_policy()
|
|
|
|
with assert_raises(S3ResponseError) as err:
|
|
bucket.get_policy()
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_website_configuration_xml():
|
|
conn = boto.connect_s3()
|
|
bucket = conn.create_bucket('test-bucket')
|
|
bucket.set_website_configuration_xml(TEST_XML)
|
|
bucket.get_website_configuration_xml().should.equal(TEST_XML)
|
|
|
|
|
|
@mock_s3_deprecated
|
|
def test_key_with_trailing_slash_in_ordinary_calling_format():
|
|
conn = boto.connect_s3(
|
|
'access_key',
|
|
'secret_key',
|
|
calling_format=boto.s3.connection.OrdinaryCallingFormat()
|
|
)
|
|
bucket = conn.create_bucket('test_bucket_name')
|
|
|
|
key_name = 'key_with_slash/'
|
|
|
|
key = Key(bucket, key_name)
|
|
key.set_contents_from_string('some value')
|
|
|
|
[k.name for k in bucket.get_all_keys()].should.contain(key_name)
|
|
|
|
|
|
"""
|
|
boto3
|
|
"""
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_key_etag():
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
s3.create_bucket(Bucket='mybucket')
|
|
s3.put_object(Bucket='mybucket', Key='steve', Body=b'is awesome')
|
|
resp = s3.get_object(Bucket='mybucket', Key='steve')
|
|
resp['ETag'].should.equal('"d32bda93738f7e03adb22e66c90fbc04"')
|
|
|
|
|
|
@mock_s3
|
|
def test_website_redirect_location():
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
s3.create_bucket(Bucket='mybucket')
|
|
|
|
s3.put_object(Bucket='mybucket', Key='steve', Body=b'is awesome')
|
|
resp = s3.get_object(Bucket='mybucket', Key='steve')
|
|
resp.get('WebsiteRedirectLocation').should.be.none
|
|
|
|
url = 'https://github.com/spulec/moto'
|
|
s3.put_object(Bucket='mybucket', Key='steve', Body=b'is awesome', WebsiteRedirectLocation=url)
|
|
resp = s3.get_object(Bucket='mybucket', Key='steve')
|
|
resp['WebsiteRedirectLocation'].should.equal(url)
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_list_keys_xml_escaped():
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
s3.create_bucket(Bucket='mybucket')
|
|
key_name = 'Q&A.txt'
|
|
s3.put_object(Bucket='mybucket', Key=key_name, Body=b'is awesome')
|
|
|
|
resp = s3.list_objects_v2(Bucket='mybucket', Prefix=key_name)
|
|
|
|
assert resp['Contents'][0]['Key'] == key_name
|
|
assert resp['KeyCount'] == 1
|
|
assert resp['MaxKeys'] == 1000
|
|
assert resp['Prefix'] == key_name
|
|
assert resp['IsTruncated'] == False
|
|
assert 'Delimiter' not in resp
|
|
assert 'StartAfter' not in resp
|
|
assert 'NextContinuationToken' not in resp
|
|
assert 'Owner' not in resp['Contents'][0]
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_list_objects_v2_truncated_response():
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
s3.create_bucket(Bucket='mybucket')
|
|
s3.put_object(Bucket='mybucket', Key='one', Body=b'1')
|
|
s3.put_object(Bucket='mybucket', Key='two', Body=b'22')
|
|
s3.put_object(Bucket='mybucket', Key='three', Body=b'333')
|
|
|
|
# First list
|
|
resp = s3.list_objects_v2(Bucket='mybucket', MaxKeys=1)
|
|
listed_object = resp['Contents'][0]
|
|
|
|
assert listed_object['Key'] == 'one'
|
|
assert resp['MaxKeys'] == 1
|
|
assert resp['Prefix'] == ''
|
|
assert resp['KeyCount'] == 1
|
|
assert resp['IsTruncated'] == True
|
|
assert 'Delimiter' not in resp
|
|
assert 'StartAfter' not in resp
|
|
assert 'Owner' not in listed_object # owner info was not requested
|
|
|
|
next_token = resp['NextContinuationToken']
|
|
|
|
# Second list
|
|
resp = s3.list_objects_v2(
|
|
Bucket='mybucket', MaxKeys=1, ContinuationToken=next_token)
|
|
listed_object = resp['Contents'][0]
|
|
|
|
assert listed_object['Key'] == 'three'
|
|
assert resp['MaxKeys'] == 1
|
|
assert resp['Prefix'] == ''
|
|
assert resp['KeyCount'] == 1
|
|
assert resp['IsTruncated'] == True
|
|
assert 'Delimiter' not in resp
|
|
assert 'StartAfter' not in resp
|
|
assert 'Owner' not in listed_object
|
|
|
|
next_token = resp['NextContinuationToken']
|
|
|
|
# Third list
|
|
resp = s3.list_objects_v2(
|
|
Bucket='mybucket', MaxKeys=1, ContinuationToken=next_token)
|
|
listed_object = resp['Contents'][0]
|
|
|
|
assert listed_object['Key'] == 'two'
|
|
assert resp['MaxKeys'] == 1
|
|
assert resp['Prefix'] == ''
|
|
assert resp['KeyCount'] == 1
|
|
assert resp['IsTruncated'] == False
|
|
assert 'Delimiter' not in resp
|
|
assert 'Owner' not in listed_object
|
|
assert 'StartAfter' not in resp
|
|
assert 'NextContinuationToken' not in resp
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_list_objects_v2_truncated_response_start_after():
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
s3.create_bucket(Bucket='mybucket')
|
|
s3.put_object(Bucket='mybucket', Key='one', Body=b'1')
|
|
s3.put_object(Bucket='mybucket', Key='two', Body=b'22')
|
|
s3.put_object(Bucket='mybucket', Key='three', Body=b'333')
|
|
|
|
# First list
|
|
resp = s3.list_objects_v2(Bucket='mybucket', MaxKeys=1, StartAfter='one')
|
|
listed_object = resp['Contents'][0]
|
|
|
|
assert listed_object['Key'] == 'three'
|
|
assert resp['MaxKeys'] == 1
|
|
assert resp['Prefix'] == ''
|
|
assert resp['KeyCount'] == 1
|
|
assert resp['IsTruncated'] == True
|
|
assert resp['StartAfter'] == 'one'
|
|
assert 'Delimiter' not in resp
|
|
assert 'Owner' not in listed_object
|
|
|
|
next_token = resp['NextContinuationToken']
|
|
|
|
# Second list
|
|
# The ContinuationToken must take precedence over StartAfter.
|
|
resp = s3.list_objects_v2(Bucket='mybucket', MaxKeys=1, StartAfter='one',
|
|
ContinuationToken=next_token)
|
|
listed_object = resp['Contents'][0]
|
|
|
|
assert listed_object['Key'] == 'two'
|
|
assert resp['MaxKeys'] == 1
|
|
assert resp['Prefix'] == ''
|
|
assert resp['KeyCount'] == 1
|
|
assert resp['IsTruncated'] == False
|
|
# When ContinuationToken is given, StartAfter is ignored. This also means
|
|
# AWS does not return it in the response.
|
|
assert 'StartAfter' not in resp
|
|
assert 'Delimiter' not in resp
|
|
assert 'Owner' not in listed_object
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_list_objects_v2_fetch_owner():
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
s3.create_bucket(Bucket='mybucket')
|
|
s3.put_object(Bucket='mybucket', Key='one', Body=b'11')
|
|
|
|
resp = s3.list_objects_v2(Bucket='mybucket', FetchOwner=True)
|
|
owner = resp['Contents'][0]['Owner']
|
|
|
|
assert 'ID' in owner
|
|
assert 'DisplayName' in owner
|
|
assert len(owner.keys()) == 2
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_bucket_create():
|
|
s3 = boto3.resource('s3', region_name='us-east-1')
|
|
s3.create_bucket(Bucket="blah")
|
|
|
|
s3.Object('blah', 'hello.txt').put(Body="some text")
|
|
|
|
s3.Object('blah', 'hello.txt').get()['Body'].read().decode(
|
|
"utf-8").should.equal("some text")
|
|
|
|
|
|
@mock_s3
|
|
def test_bucket_create_duplicate():
|
|
s3 = boto3.resource('s3', region_name='us-west-2')
|
|
s3.create_bucket(Bucket="blah", CreateBucketConfiguration={
|
|
'LocationConstraint': 'us-west-2',
|
|
})
|
|
with assert_raises(ClientError) as exc:
|
|
s3.create_bucket(
|
|
Bucket="blah",
|
|
CreateBucketConfiguration={
|
|
'LocationConstraint': 'us-west-2',
|
|
}
|
|
)
|
|
exc.exception.response['Error']['Code'].should.equal('BucketAlreadyExists')
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_bucket_create_eu_central():
|
|
s3 = boto3.resource('s3', region_name='eu-central-1')
|
|
s3.create_bucket(Bucket="blah")
|
|
|
|
s3.Object('blah', 'hello.txt').put(Body="some text")
|
|
|
|
s3.Object('blah', 'hello.txt').get()['Body'].read().decode(
|
|
"utf-8").should.equal("some text")
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_head_object():
|
|
s3 = boto3.resource('s3', region_name='us-east-1')
|
|
s3.create_bucket(Bucket="blah")
|
|
|
|
s3.Object('blah', 'hello.txt').put(Body="some text")
|
|
|
|
s3.Object('blah', 'hello.txt').meta.client.head_object(
|
|
Bucket='blah', Key='hello.txt')
|
|
|
|
with assert_raises(ClientError) as e:
|
|
s3.Object('blah', 'hello2.txt').meta.client.head_object(
|
|
Bucket='blah', Key='hello_bad.txt')
|
|
e.exception.response['Error']['Code'].should.equal('404')
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_bucket_deletion():
|
|
cli = boto3.client('s3', region_name='us-east-1')
|
|
cli.create_bucket(Bucket="foobar")
|
|
|
|
cli.put_object(Bucket="foobar", Key="the-key", Body="some value")
|
|
|
|
# Try to delete a bucket that still has keys
|
|
cli.delete_bucket.when.called_with(Bucket="foobar").should.throw(
|
|
cli.exceptions.ClientError,
|
|
('An error occurred (BucketNotEmpty) when calling the DeleteBucket operation: '
|
|
'The bucket you tried to delete is not empty'))
|
|
|
|
cli.delete_object(Bucket="foobar", Key="the-key")
|
|
cli.delete_bucket(Bucket="foobar")
|
|
|
|
# Get non-existing bucket
|
|
cli.head_bucket.when.called_with(Bucket="foobar").should.throw(
|
|
cli.exceptions.ClientError,
|
|
"An error occurred (404) when calling the HeadBucket operation: Not Found")
|
|
|
|
# Delete non-existing bucket
|
|
cli.delete_bucket.when.called_with(Bucket="foobar").should.throw(cli.exceptions.NoSuchBucket)
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_get_object():
|
|
s3 = boto3.resource('s3', region_name='us-east-1')
|
|
s3.create_bucket(Bucket="blah")
|
|
|
|
s3.Object('blah', 'hello.txt').put(Body="some text")
|
|
|
|
s3.Object('blah', 'hello.txt').meta.client.head_object(
|
|
Bucket='blah', Key='hello.txt')
|
|
|
|
with assert_raises(ClientError) as e:
|
|
s3.Object('blah', 'hello2.txt').get()
|
|
|
|
e.exception.response['Error']['Code'].should.equal('NoSuchKey')
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_head_object_with_versioning():
|
|
s3 = boto3.resource('s3', region_name='us-east-1')
|
|
bucket = s3.create_bucket(Bucket='blah')
|
|
bucket.Versioning().enable()
|
|
|
|
old_content = 'some text'
|
|
new_content = 'some new text'
|
|
s3.Object('blah', 'hello.txt').put(Body=old_content)
|
|
s3.Object('blah', 'hello.txt').put(Body=new_content)
|
|
|
|
head_object = s3.Object('blah', 'hello.txt').meta.client.head_object(
|
|
Bucket='blah', Key='hello.txt')
|
|
head_object['VersionId'].should.equal('1')
|
|
head_object['ContentLength'].should.equal(len(new_content))
|
|
|
|
old_head_object = s3.Object('blah', 'hello.txt').meta.client.head_object(
|
|
Bucket='blah', Key='hello.txt', VersionId='0')
|
|
old_head_object['VersionId'].should.equal('0')
|
|
old_head_object['ContentLength'].should.equal(len(old_content))
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_copy_object_with_versioning():
|
|
client = boto3.client('s3', region_name='us-east-1')
|
|
|
|
client.create_bucket(Bucket='blah', CreateBucketConfiguration={'LocationConstraint': 'eu-west-1'})
|
|
client.put_bucket_versioning(Bucket='blah', VersioningConfiguration={'Status': 'Enabled'})
|
|
|
|
client.put_object(Bucket='blah', Key='test1', Body=b'test1')
|
|
client.put_object(Bucket='blah', Key='test2', Body=b'test2')
|
|
|
|
obj1_version = client.get_object(Bucket='blah', Key='test1')['VersionId']
|
|
obj2_version = client.get_object(Bucket='blah', Key='test2')['VersionId']
|
|
|
|
# Versions should be the same
|
|
obj1_version.should.equal(obj2_version)
|
|
|
|
client.copy_object(CopySource={'Bucket': 'blah', 'Key': 'test1'}, Bucket='blah', Key='test2')
|
|
obj2_version_new = client.get_object(Bucket='blah', Key='test2')['VersionId']
|
|
|
|
# Version should be different to previous version
|
|
obj2_version_new.should_not.equal(obj2_version)
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_deleted_versionings_list():
|
|
client = boto3.client('s3', region_name='us-east-1')
|
|
|
|
client.create_bucket(Bucket='blah')
|
|
client.put_bucket_versioning(Bucket='blah', VersioningConfiguration={'Status': 'Enabled'})
|
|
|
|
client.put_object(Bucket='blah', Key='test1', Body=b'test1')
|
|
client.put_object(Bucket='blah', Key='test2', Body=b'test2')
|
|
client.delete_objects(Bucket='blah', Delete={'Objects': [{'Key': 'test1'}]})
|
|
|
|
listed = client.list_objects_v2(Bucket='blah')
|
|
assert len(listed['Contents']) == 1
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_delete_versioned_bucket():
|
|
client = boto3.client('s3', region_name='us-east-1')
|
|
|
|
client.create_bucket(Bucket='blah')
|
|
client.put_bucket_versioning(Bucket='blah', VersioningConfiguration={'Status': 'Enabled'})
|
|
|
|
resp = client.put_object(Bucket='blah', Key='test1', Body=b'test1')
|
|
client.delete_object(Bucket='blah', Key='test1', VersionId=resp["VersionId"])
|
|
|
|
client.delete_bucket(Bucket='blah')
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_head_object_if_modified_since():
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
bucket_name = "blah"
|
|
s3.create_bucket(Bucket=bucket_name)
|
|
|
|
key = 'hello.txt'
|
|
|
|
s3.put_object(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
Body='test'
|
|
)
|
|
|
|
with assert_raises(botocore.exceptions.ClientError) as err:
|
|
s3.head_object(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
IfModifiedSince=datetime.datetime.utcnow() + datetime.timedelta(hours=1)
|
|
)
|
|
e = err.exception
|
|
e.response['Error'].should.equal({'Code': '304', 'Message': 'Not Modified'})
|
|
|
|
|
|
@mock_s3
|
|
@reduced_min_part_size
|
|
def test_boto3_multipart_etag():
|
|
# Create Bucket so that test can run
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
s3.create_bucket(Bucket='mybucket')
|
|
|
|
upload_id = s3.create_multipart_upload(
|
|
Bucket='mybucket', Key='the-key')['UploadId']
|
|
part1 = b'0' * REDUCED_PART_SIZE
|
|
etags = []
|
|
etags.append(
|
|
s3.upload_part(Bucket='mybucket', Key='the-key', PartNumber=1,
|
|
UploadId=upload_id, Body=part1)['ETag'])
|
|
# last part, can be less than 5 MB
|
|
part2 = b'1'
|
|
etags.append(
|
|
s3.upload_part(Bucket='mybucket', Key='the-key', PartNumber=2,
|
|
UploadId=upload_id, Body=part2)['ETag'])
|
|
s3.complete_multipart_upload(
|
|
Bucket='mybucket', Key='the-key', UploadId=upload_id,
|
|
MultipartUpload={'Parts': [{'ETag': etag, 'PartNumber': i}
|
|
for i, etag in enumerate(etags, 1)]})
|
|
# we should get both parts as the key contents
|
|
resp = s3.get_object(Bucket='mybucket', Key='the-key')
|
|
resp['ETag'].should.equal(EXPECTED_ETAG)
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_put_object_with_tagging():
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
bucket_name = 'mybucket'
|
|
key = 'key-with-tags'
|
|
s3.create_bucket(Bucket=bucket_name)
|
|
|
|
s3.put_object(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
Body='test',
|
|
Tagging='foo=bar',
|
|
)
|
|
|
|
resp = s3.get_object_tagging(Bucket=bucket_name, Key=key)
|
|
|
|
resp['TagSet'].should.contain({'Key': 'foo', 'Value': 'bar'})
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_put_bucket_tagging():
|
|
s3 = boto3.client("s3", region_name="us-east-1")
|
|
bucket_name = "mybucket"
|
|
s3.create_bucket(Bucket=bucket_name)
|
|
|
|
# With 1 tag:
|
|
resp = s3.put_bucket_tagging(Bucket=bucket_name,
|
|
Tagging={
|
|
"TagSet": [
|
|
{
|
|
"Key": "TagOne",
|
|
"Value": "ValueOne"
|
|
}
|
|
]
|
|
})
|
|
resp['ResponseMetadata']['HTTPStatusCode'].should.equal(200)
|
|
|
|
# With multiple tags:
|
|
resp = s3.put_bucket_tagging(Bucket=bucket_name,
|
|
Tagging={
|
|
"TagSet": [
|
|
{
|
|
"Key": "TagOne",
|
|
"Value": "ValueOne"
|
|
},
|
|
{
|
|
"Key": "TagTwo",
|
|
"Value": "ValueTwo"
|
|
}
|
|
]
|
|
})
|
|
|
|
resp['ResponseMetadata']['HTTPStatusCode'].should.equal(200)
|
|
|
|
# No tags is also OK:
|
|
resp = s3.put_bucket_tagging(Bucket=bucket_name, Tagging={
|
|
"TagSet": []
|
|
})
|
|
resp['ResponseMetadata']['HTTPStatusCode'].should.equal(200)
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_get_bucket_tagging():
|
|
s3 = boto3.client("s3", region_name="us-east-1")
|
|
bucket_name = "mybucket"
|
|
s3.create_bucket(Bucket=bucket_name)
|
|
s3.put_bucket_tagging(Bucket=bucket_name,
|
|
Tagging={
|
|
"TagSet": [
|
|
{
|
|
"Key": "TagOne",
|
|
"Value": "ValueOne"
|
|
},
|
|
{
|
|
"Key": "TagTwo",
|
|
"Value": "ValueTwo"
|
|
}
|
|
]
|
|
})
|
|
|
|
# Get the tags for the bucket:
|
|
resp = s3.get_bucket_tagging(Bucket=bucket_name)
|
|
resp['ResponseMetadata']['HTTPStatusCode'].should.equal(200)
|
|
len(resp["TagSet"]).should.equal(2)
|
|
|
|
# With no tags:
|
|
s3.put_bucket_tagging(Bucket=bucket_name, Tagging={
|
|
"TagSet": []
|
|
})
|
|
|
|
with assert_raises(ClientError) as err:
|
|
s3.get_bucket_tagging(Bucket=bucket_name)
|
|
|
|
e = err.exception
|
|
e.response["Error"]["Code"].should.equal("NoSuchTagSet")
|
|
e.response["Error"]["Message"].should.equal("The TagSet does not exist")
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_delete_bucket_tagging():
|
|
s3 = boto3.client("s3", region_name="us-east-1")
|
|
bucket_name = "mybucket"
|
|
s3.create_bucket(Bucket=bucket_name)
|
|
|
|
s3.put_bucket_tagging(Bucket=bucket_name,
|
|
Tagging={
|
|
"TagSet": [
|
|
{
|
|
"Key": "TagOne",
|
|
"Value": "ValueOne"
|
|
},
|
|
{
|
|
"Key": "TagTwo",
|
|
"Value": "ValueTwo"
|
|
}
|
|
]
|
|
})
|
|
|
|
resp = s3.delete_bucket_tagging(Bucket=bucket_name)
|
|
resp['ResponseMetadata']['HTTPStatusCode'].should.equal(204)
|
|
|
|
with assert_raises(ClientError) as err:
|
|
s3.get_bucket_tagging(Bucket=bucket_name)
|
|
|
|
e = err.exception
|
|
e.response["Error"]["Code"].should.equal("NoSuchTagSet")
|
|
e.response["Error"]["Message"].should.equal("The TagSet does not exist")
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_put_bucket_cors():
|
|
s3 = boto3.client("s3", region_name="us-east-1")
|
|
bucket_name = "mybucket"
|
|
s3.create_bucket(Bucket=bucket_name)
|
|
|
|
resp = s3.put_bucket_cors(Bucket=bucket_name, CORSConfiguration={
|
|
"CORSRules": [
|
|
{
|
|
"AllowedOrigins": [
|
|
"*"
|
|
],
|
|
"AllowedMethods": [
|
|
"GET",
|
|
"POST"
|
|
],
|
|
"AllowedHeaders": [
|
|
"Authorization"
|
|
],
|
|
"ExposeHeaders": [
|
|
"x-amz-request-id"
|
|
],
|
|
"MaxAgeSeconds": 123
|
|
},
|
|
{
|
|
"AllowedOrigins": [
|
|
"*"
|
|
],
|
|
"AllowedMethods": [
|
|
"PUT"
|
|
],
|
|
"AllowedHeaders": [
|
|
"Authorization"
|
|
],
|
|
"ExposeHeaders": [
|
|
"x-amz-request-id"
|
|
],
|
|
"MaxAgeSeconds": 123
|
|
}
|
|
]
|
|
})
|
|
|
|
resp['ResponseMetadata']['HTTPStatusCode'].should.equal(200)
|
|
|
|
with assert_raises(ClientError) as err:
|
|
s3.put_bucket_cors(Bucket=bucket_name, CORSConfiguration={
|
|
"CORSRules": [
|
|
{
|
|
"AllowedOrigins": [
|
|
"*"
|
|
],
|
|
"AllowedMethods": [
|
|
"NOTREAL",
|
|
"POST"
|
|
]
|
|
}
|
|
]
|
|
})
|
|
e = err.exception
|
|
e.response["Error"]["Code"].should.equal("InvalidRequest")
|
|
e.response["Error"]["Message"].should.equal("Found unsupported HTTP method in CORS config. "
|
|
"Unsupported method is NOTREAL")
|
|
|
|
with assert_raises(ClientError) as err:
|
|
s3.put_bucket_cors(Bucket=bucket_name, CORSConfiguration={
|
|
"CORSRules": []
|
|
})
|
|
e = err.exception
|
|
e.response["Error"]["Code"].should.equal("MalformedXML")
|
|
|
|
# And 101:
|
|
many_rules = [{"AllowedOrigins": ["*"], "AllowedMethods": ["GET"]}] * 101
|
|
with assert_raises(ClientError) as err:
|
|
s3.put_bucket_cors(Bucket=bucket_name, CORSConfiguration={
|
|
"CORSRules": many_rules
|
|
})
|
|
e = err.exception
|
|
e.response["Error"]["Code"].should.equal("MalformedXML")
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_get_bucket_cors():
|
|
s3 = boto3.client("s3", region_name="us-east-1")
|
|
bucket_name = "mybucket"
|
|
s3.create_bucket(Bucket=bucket_name)
|
|
|
|
# Without CORS:
|
|
with assert_raises(ClientError) as err:
|
|
s3.get_bucket_cors(Bucket=bucket_name)
|
|
|
|
e = err.exception
|
|
e.response["Error"]["Code"].should.equal("NoSuchCORSConfiguration")
|
|
e.response["Error"]["Message"].should.equal("The CORS configuration does not exist")
|
|
|
|
s3.put_bucket_cors(Bucket=bucket_name, CORSConfiguration={
|
|
"CORSRules": [
|
|
{
|
|
"AllowedOrigins": [
|
|
"*"
|
|
],
|
|
"AllowedMethods": [
|
|
"GET",
|
|
"POST"
|
|
],
|
|
"AllowedHeaders": [
|
|
"Authorization"
|
|
],
|
|
"ExposeHeaders": [
|
|
"x-amz-request-id"
|
|
],
|
|
"MaxAgeSeconds": 123
|
|
},
|
|
{
|
|
"AllowedOrigins": [
|
|
"*"
|
|
],
|
|
"AllowedMethods": [
|
|
"PUT"
|
|
],
|
|
"AllowedHeaders": [
|
|
"Authorization"
|
|
],
|
|
"ExposeHeaders": [
|
|
"x-amz-request-id"
|
|
],
|
|
"MaxAgeSeconds": 123
|
|
}
|
|
]
|
|
})
|
|
|
|
resp = s3.get_bucket_cors(Bucket=bucket_name)
|
|
resp['ResponseMetadata']['HTTPStatusCode'].should.equal(200)
|
|
len(resp["CORSRules"]).should.equal(2)
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_delete_bucket_cors():
|
|
s3 = boto3.client("s3", region_name="us-east-1")
|
|
bucket_name = "mybucket"
|
|
s3.create_bucket(Bucket=bucket_name)
|
|
s3.put_bucket_cors(Bucket=bucket_name, CORSConfiguration={
|
|
"CORSRules": [
|
|
{
|
|
"AllowedOrigins": [
|
|
"*"
|
|
],
|
|
"AllowedMethods": [
|
|
"GET"
|
|
]
|
|
}
|
|
]
|
|
})
|
|
|
|
resp = s3.delete_bucket_cors(Bucket=bucket_name)
|
|
resp['ResponseMetadata']['HTTPStatusCode'].should.equal(204)
|
|
|
|
# Verify deletion:
|
|
with assert_raises(ClientError) as err:
|
|
s3.get_bucket_cors(Bucket=bucket_name)
|
|
|
|
e = err.exception
|
|
e.response["Error"]["Code"].should.equal("NoSuchCORSConfiguration")
|
|
e.response["Error"]["Message"].should.equal("The CORS configuration does not exist")
|
|
|
|
|
|
@mock_s3
|
|
def test_put_bucket_acl_body():
|
|
s3 = boto3.client("s3", region_name="us-east-1")
|
|
s3.create_bucket(Bucket="bucket")
|
|
bucket_owner = s3.get_bucket_acl(Bucket="bucket")["Owner"]
|
|
s3.put_bucket_acl(Bucket="bucket", AccessControlPolicy={
|
|
"Grants": [
|
|
{
|
|
"Grantee": {
|
|
"URI": "http://acs.amazonaws.com/groups/s3/LogDelivery",
|
|
"Type": "Group"
|
|
},
|
|
"Permission": "WRITE"
|
|
},
|
|
{
|
|
"Grantee": {
|
|
"URI": "http://acs.amazonaws.com/groups/s3/LogDelivery",
|
|
"Type": "Group"
|
|
},
|
|
"Permission": "READ_ACP"
|
|
}
|
|
],
|
|
"Owner": bucket_owner
|
|
})
|
|
|
|
result = s3.get_bucket_acl(Bucket="bucket")
|
|
assert len(result["Grants"]) == 2
|
|
for g in result["Grants"]:
|
|
assert g["Grantee"]["URI"] == "http://acs.amazonaws.com/groups/s3/LogDelivery"
|
|
assert g["Grantee"]["Type"] == "Group"
|
|
assert g["Permission"] in ["WRITE", "READ_ACP"]
|
|
|
|
# With one:
|
|
s3.put_bucket_acl(Bucket="bucket", AccessControlPolicy={
|
|
"Grants": [
|
|
{
|
|
"Grantee": {
|
|
"URI": "http://acs.amazonaws.com/groups/s3/LogDelivery",
|
|
"Type": "Group"
|
|
},
|
|
"Permission": "WRITE"
|
|
}
|
|
],
|
|
"Owner": bucket_owner
|
|
})
|
|
result = s3.get_bucket_acl(Bucket="bucket")
|
|
assert len(result["Grants"]) == 1
|
|
|
|
# With no owner:
|
|
with assert_raises(ClientError) as err:
|
|
s3.put_bucket_acl(Bucket="bucket", AccessControlPolicy={
|
|
"Grants": [
|
|
{
|
|
"Grantee": {
|
|
"URI": "http://acs.amazonaws.com/groups/s3/LogDelivery",
|
|
"Type": "Group"
|
|
},
|
|
"Permission": "WRITE"
|
|
}
|
|
]
|
|
})
|
|
assert err.exception.response["Error"]["Code"] == "MalformedACLError"
|
|
|
|
# With incorrect permission:
|
|
with assert_raises(ClientError) as err:
|
|
s3.put_bucket_acl(Bucket="bucket", AccessControlPolicy={
|
|
"Grants": [
|
|
{
|
|
"Grantee": {
|
|
"URI": "http://acs.amazonaws.com/groups/s3/LogDelivery",
|
|
"Type": "Group"
|
|
},
|
|
"Permission": "lskjflkasdjflkdsjfalisdjflkdsjf"
|
|
}
|
|
],
|
|
"Owner": bucket_owner
|
|
})
|
|
assert err.exception.response["Error"]["Code"] == "MalformedACLError"
|
|
|
|
# Clear the ACLs:
|
|
result = s3.put_bucket_acl(Bucket="bucket", AccessControlPolicy={"Grants": [], "Owner": bucket_owner})
|
|
assert not result.get("Grants")
|
|
|
|
|
|
@mock_s3
|
|
def test_put_bucket_notification():
|
|
s3 = boto3.client("s3", region_name="us-east-1")
|
|
s3.create_bucket(Bucket="bucket")
|
|
|
|
# With no configuration:
|
|
result = s3.get_bucket_notification(Bucket="bucket")
|
|
assert not result.get("TopicConfigurations")
|
|
assert not result.get("QueueConfigurations")
|
|
assert not result.get("LambdaFunctionConfigurations")
|
|
|
|
# Place proper topic configuration:
|
|
s3.put_bucket_notification_configuration(Bucket="bucket",
|
|
NotificationConfiguration={
|
|
"TopicConfigurations": [
|
|
{
|
|
"TopicArn": "arn:aws:sns:us-east-1:012345678910:mytopic",
|
|
"Events": [
|
|
"s3:ObjectCreated:*",
|
|
"s3:ObjectRemoved:*"
|
|
]
|
|
},
|
|
{
|
|
"TopicArn": "arn:aws:sns:us-east-1:012345678910:myothertopic",
|
|
"Events": [
|
|
"s3:ObjectCreated:*"
|
|
],
|
|
"Filter": {
|
|
"Key": {
|
|
"FilterRules": [
|
|
{
|
|
"Name": "prefix",
|
|
"Value": "images/"
|
|
},
|
|
{
|
|
"Name": "suffix",
|
|
"Value": "png"
|
|
}
|
|
]
|
|
}
|
|
}
|
|
}
|
|
]
|
|
})
|
|
|
|
# Verify to completion:
|
|
result = s3.get_bucket_notification_configuration(Bucket="bucket")
|
|
assert len(result["TopicConfigurations"]) == 2
|
|
assert not result.get("QueueConfigurations")
|
|
assert not result.get("LambdaFunctionConfigurations")
|
|
assert result["TopicConfigurations"][0]["TopicArn"] == "arn:aws:sns:us-east-1:012345678910:mytopic"
|
|
assert result["TopicConfigurations"][1]["TopicArn"] == "arn:aws:sns:us-east-1:012345678910:myothertopic"
|
|
assert len(result["TopicConfigurations"][0]["Events"]) == 2
|
|
assert len(result["TopicConfigurations"][1]["Events"]) == 1
|
|
assert result["TopicConfigurations"][0]["Events"][0] == "s3:ObjectCreated:*"
|
|
assert result["TopicConfigurations"][0]["Events"][1] == "s3:ObjectRemoved:*"
|
|
assert result["TopicConfigurations"][1]["Events"][0] == "s3:ObjectCreated:*"
|
|
assert result["TopicConfigurations"][0]["Id"]
|
|
assert result["TopicConfigurations"][1]["Id"]
|
|
assert not result["TopicConfigurations"][0].get("Filter")
|
|
assert len(result["TopicConfigurations"][1]["Filter"]["Key"]["FilterRules"]) == 2
|
|
assert result["TopicConfigurations"][1]["Filter"]["Key"]["FilterRules"][0]["Name"] == "prefix"
|
|
assert result["TopicConfigurations"][1]["Filter"]["Key"]["FilterRules"][0]["Value"] == "images/"
|
|
assert result["TopicConfigurations"][1]["Filter"]["Key"]["FilterRules"][1]["Name"] == "suffix"
|
|
assert result["TopicConfigurations"][1]["Filter"]["Key"]["FilterRules"][1]["Value"] == "png"
|
|
|
|
# Place proper queue configuration:
|
|
s3.put_bucket_notification_configuration(Bucket="bucket",
|
|
NotificationConfiguration={
|
|
"QueueConfigurations": [
|
|
{
|
|
"Id": "SomeID",
|
|
"QueueArn": "arn:aws:sqs:us-east-1:012345678910:myQueue",
|
|
"Events": ["s3:ObjectCreated:*"],
|
|
"Filter": {
|
|
"Key": {
|
|
"FilterRules": [
|
|
{
|
|
"Name": "prefix",
|
|
"Value": "images/"
|
|
}
|
|
]
|
|
}
|
|
}
|
|
}
|
|
]
|
|
})
|
|
result = s3.get_bucket_notification_configuration(Bucket="bucket")
|
|
assert len(result["QueueConfigurations"]) == 1
|
|
assert not result.get("TopicConfigurations")
|
|
assert not result.get("LambdaFunctionConfigurations")
|
|
assert result["QueueConfigurations"][0]["Id"] == "SomeID"
|
|
assert result["QueueConfigurations"][0]["QueueArn"] == "arn:aws:sqs:us-east-1:012345678910:myQueue"
|
|
assert result["QueueConfigurations"][0]["Events"][0] == "s3:ObjectCreated:*"
|
|
assert len(result["QueueConfigurations"][0]["Events"]) == 1
|
|
assert len(result["QueueConfigurations"][0]["Filter"]["Key"]["FilterRules"]) == 1
|
|
assert result["QueueConfigurations"][0]["Filter"]["Key"]["FilterRules"][0]["Name"] == "prefix"
|
|
assert result["QueueConfigurations"][0]["Filter"]["Key"]["FilterRules"][0]["Value"] == "images/"
|
|
|
|
# Place proper Lambda configuration:
|
|
s3.put_bucket_notification_configuration(Bucket="bucket",
|
|
NotificationConfiguration={
|
|
"LambdaFunctionConfigurations": [
|
|
{
|
|
"LambdaFunctionArn":
|
|
"arn:aws:lambda:us-east-1:012345678910:function:lambda",
|
|
"Events": ["s3:ObjectCreated:*"],
|
|
"Filter": {
|
|
"Key": {
|
|
"FilterRules": [
|
|
{
|
|
"Name": "prefix",
|
|
"Value": "images/"
|
|
}
|
|
]
|
|
}
|
|
}
|
|
}
|
|
]
|
|
})
|
|
result = s3.get_bucket_notification_configuration(Bucket="bucket")
|
|
assert len(result["LambdaFunctionConfigurations"]) == 1
|
|
assert not result.get("TopicConfigurations")
|
|
assert not result.get("QueueConfigurations")
|
|
assert result["LambdaFunctionConfigurations"][0]["Id"]
|
|
assert result["LambdaFunctionConfigurations"][0]["LambdaFunctionArn"] == \
|
|
"arn:aws:lambda:us-east-1:012345678910:function:lambda"
|
|
assert result["LambdaFunctionConfigurations"][0]["Events"][0] == "s3:ObjectCreated:*"
|
|
assert len(result["LambdaFunctionConfigurations"][0]["Events"]) == 1
|
|
assert len(result["LambdaFunctionConfigurations"][0]["Filter"]["Key"]["FilterRules"]) == 1
|
|
assert result["LambdaFunctionConfigurations"][0]["Filter"]["Key"]["FilterRules"][0]["Name"] == "prefix"
|
|
assert result["LambdaFunctionConfigurations"][0]["Filter"]["Key"]["FilterRules"][0]["Value"] == "images/"
|
|
|
|
# And with all 3 set:
|
|
s3.put_bucket_notification_configuration(Bucket="bucket",
|
|
NotificationConfiguration={
|
|
"TopicConfigurations": [
|
|
{
|
|
"TopicArn": "arn:aws:sns:us-east-1:012345678910:mytopic",
|
|
"Events": [
|
|
"s3:ObjectCreated:*",
|
|
"s3:ObjectRemoved:*"
|
|
]
|
|
}
|
|
],
|
|
"LambdaFunctionConfigurations": [
|
|
{
|
|
"LambdaFunctionArn":
|
|
"arn:aws:lambda:us-east-1:012345678910:function:lambda",
|
|
"Events": ["s3:ObjectCreated:*"]
|
|
}
|
|
],
|
|
"QueueConfigurations": [
|
|
{
|
|
"QueueArn": "arn:aws:sqs:us-east-1:012345678910:myQueue",
|
|
"Events": ["s3:ObjectCreated:*"]
|
|
}
|
|
]
|
|
})
|
|
result = s3.get_bucket_notification_configuration(Bucket="bucket")
|
|
assert len(result["LambdaFunctionConfigurations"]) == 1
|
|
assert len(result["TopicConfigurations"]) == 1
|
|
assert len(result["QueueConfigurations"]) == 1
|
|
|
|
# And clear it out:
|
|
s3.put_bucket_notification_configuration(Bucket="bucket", NotificationConfiguration={})
|
|
result = s3.get_bucket_notification_configuration(Bucket="bucket")
|
|
assert not result.get("TopicConfigurations")
|
|
assert not result.get("QueueConfigurations")
|
|
assert not result.get("LambdaFunctionConfigurations")
|
|
|
|
|
|
@mock_s3
|
|
def test_put_bucket_notification_errors():
|
|
s3 = boto3.client("s3", region_name="us-east-1")
|
|
s3.create_bucket(Bucket="bucket")
|
|
|
|
# With incorrect ARNs:
|
|
for tech, arn in [("Queue", "sqs"), ("Topic", "sns"), ("LambdaFunction", "lambda")]:
|
|
with assert_raises(ClientError) as err:
|
|
s3.put_bucket_notification_configuration(Bucket="bucket",
|
|
NotificationConfiguration={
|
|
"{}Configurations".format(tech): [
|
|
{
|
|
"{}Arn".format(tech):
|
|
"arn:aws:{}:us-east-1:012345678910:lksajdfkldskfj",
|
|
"Events": ["s3:ObjectCreated:*"]
|
|
}
|
|
]
|
|
})
|
|
|
|
assert err.exception.response["Error"]["Code"] == "InvalidArgument"
|
|
assert err.exception.response["Error"]["Message"] == "The ARN is not well formed"
|
|
|
|
# Region not the same as the bucket:
|
|
with assert_raises(ClientError) as err:
|
|
s3.put_bucket_notification_configuration(Bucket="bucket",
|
|
NotificationConfiguration={
|
|
"QueueConfigurations": [
|
|
{
|
|
"QueueArn":
|
|
"arn:aws:sqs:us-west-2:012345678910:lksajdfkldskfj",
|
|
"Events": ["s3:ObjectCreated:*"]
|
|
}
|
|
]
|
|
})
|
|
|
|
assert err.exception.response["Error"]["Code"] == "InvalidArgument"
|
|
assert err.exception.response["Error"]["Message"] == \
|
|
"The notification destination service region is not valid for the bucket location constraint"
|
|
|
|
# Invalid event name:
|
|
with assert_raises(ClientError) as err:
|
|
s3.put_bucket_notification_configuration(Bucket="bucket",
|
|
NotificationConfiguration={
|
|
"QueueConfigurations": [
|
|
{
|
|
"QueueArn":
|
|
"arn:aws:sqs:us-east-1:012345678910:lksajdfkldskfj",
|
|
"Events": ["notarealeventname"]
|
|
}
|
|
]
|
|
})
|
|
assert err.exception.response["Error"]["Code"] == "InvalidArgument"
|
|
assert err.exception.response["Error"]["Message"] == "The event is not supported for notifications"
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_put_bucket_logging():
|
|
s3 = boto3.client("s3", region_name="us-east-1")
|
|
bucket_name = "mybucket"
|
|
log_bucket = "logbucket"
|
|
wrong_region_bucket = "wrongregionlogbucket"
|
|
s3.create_bucket(Bucket=bucket_name)
|
|
s3.create_bucket(Bucket=log_bucket) # Adding the ACL for log-delivery later...
|
|
s3.create_bucket(Bucket=wrong_region_bucket, CreateBucketConfiguration={"LocationConstraint": "us-west-2"})
|
|
|
|
# No logging config:
|
|
result = s3.get_bucket_logging(Bucket=bucket_name)
|
|
assert not result.get("LoggingEnabled")
|
|
|
|
# A log-bucket that doesn't exist:
|
|
with assert_raises(ClientError) as err:
|
|
s3.put_bucket_logging(Bucket=bucket_name, BucketLoggingStatus={
|
|
"LoggingEnabled": {
|
|
"TargetBucket": "IAMNOTREAL",
|
|
"TargetPrefix": ""
|
|
}
|
|
})
|
|
assert err.exception.response["Error"]["Code"] == "InvalidTargetBucketForLogging"
|
|
|
|
# A log-bucket that's missing the proper ACLs for LogDelivery:
|
|
with assert_raises(ClientError) as err:
|
|
s3.put_bucket_logging(Bucket=bucket_name, BucketLoggingStatus={
|
|
"LoggingEnabled": {
|
|
"TargetBucket": log_bucket,
|
|
"TargetPrefix": ""
|
|
}
|
|
})
|
|
assert err.exception.response["Error"]["Code"] == "InvalidTargetBucketForLogging"
|
|
assert "log-delivery" in err.exception.response["Error"]["Message"]
|
|
|
|
# Add the proper "log-delivery" ACL to the log buckets:
|
|
bucket_owner = s3.get_bucket_acl(Bucket=log_bucket)["Owner"]
|
|
for bucket in [log_bucket, wrong_region_bucket]:
|
|
s3.put_bucket_acl(Bucket=bucket, AccessControlPolicy={
|
|
"Grants": [
|
|
{
|
|
"Grantee": {
|
|
"URI": "http://acs.amazonaws.com/groups/s3/LogDelivery",
|
|
"Type": "Group"
|
|
},
|
|
"Permission": "WRITE"
|
|
},
|
|
{
|
|
"Grantee": {
|
|
"URI": "http://acs.amazonaws.com/groups/s3/LogDelivery",
|
|
"Type": "Group"
|
|
},
|
|
"Permission": "READ_ACP"
|
|
},
|
|
{
|
|
"Grantee": {
|
|
"Type": "CanonicalUser",
|
|
"ID": bucket_owner["ID"]
|
|
},
|
|
"Permission": "FULL_CONTROL"
|
|
}
|
|
],
|
|
"Owner": bucket_owner
|
|
})
|
|
|
|
# A log-bucket that's in the wrong region:
|
|
with assert_raises(ClientError) as err:
|
|
s3.put_bucket_logging(Bucket=bucket_name, BucketLoggingStatus={
|
|
"LoggingEnabled": {
|
|
"TargetBucket": wrong_region_bucket,
|
|
"TargetPrefix": ""
|
|
}
|
|
})
|
|
assert err.exception.response["Error"]["Code"] == "CrossLocationLoggingProhibitted"
|
|
|
|
# Correct logging:
|
|
s3.put_bucket_logging(Bucket=bucket_name, BucketLoggingStatus={
|
|
"LoggingEnabled": {
|
|
"TargetBucket": log_bucket,
|
|
"TargetPrefix": "{}/".format(bucket_name)
|
|
}
|
|
})
|
|
result = s3.get_bucket_logging(Bucket=bucket_name)
|
|
assert result["LoggingEnabled"]["TargetBucket"] == log_bucket
|
|
assert result["LoggingEnabled"]["TargetPrefix"] == "{}/".format(bucket_name)
|
|
assert not result["LoggingEnabled"].get("TargetGrants")
|
|
|
|
# And disabling:
|
|
s3.put_bucket_logging(Bucket=bucket_name, BucketLoggingStatus={})
|
|
assert not s3.get_bucket_logging(Bucket=bucket_name).get("LoggingEnabled")
|
|
|
|
# And enabling with multiple target grants:
|
|
s3.put_bucket_logging(Bucket=bucket_name, BucketLoggingStatus={
|
|
"LoggingEnabled": {
|
|
"TargetBucket": log_bucket,
|
|
"TargetPrefix": "{}/".format(bucket_name),
|
|
"TargetGrants": [
|
|
{
|
|
"Grantee": {
|
|
"ID": "SOMEIDSTRINGHERE9238748923734823917498237489237409123840983274",
|
|
"Type": "CanonicalUser"
|
|
},
|
|
"Permission": "READ"
|
|
},
|
|
{
|
|
"Grantee": {
|
|
"ID": "SOMEIDSTRINGHERE9238748923734823917498237489237409123840983274",
|
|
"Type": "CanonicalUser"
|
|
},
|
|
"Permission": "WRITE"
|
|
}
|
|
]
|
|
}
|
|
})
|
|
|
|
result = s3.get_bucket_logging(Bucket=bucket_name)
|
|
assert len(result["LoggingEnabled"]["TargetGrants"]) == 2
|
|
assert result["LoggingEnabled"]["TargetGrants"][0]["Grantee"]["ID"] == \
|
|
"SOMEIDSTRINGHERE9238748923734823917498237489237409123840983274"
|
|
|
|
# Test with just 1 grant:
|
|
s3.put_bucket_logging(Bucket=bucket_name, BucketLoggingStatus={
|
|
"LoggingEnabled": {
|
|
"TargetBucket": log_bucket,
|
|
"TargetPrefix": "{}/".format(bucket_name),
|
|
"TargetGrants": [
|
|
{
|
|
"Grantee": {
|
|
"ID": "SOMEIDSTRINGHERE9238748923734823917498237489237409123840983274",
|
|
"Type": "CanonicalUser"
|
|
},
|
|
"Permission": "READ"
|
|
}
|
|
]
|
|
}
|
|
})
|
|
result = s3.get_bucket_logging(Bucket=bucket_name)
|
|
assert len(result["LoggingEnabled"]["TargetGrants"]) == 1
|
|
|
|
# With an invalid grant:
|
|
with assert_raises(ClientError) as err:
|
|
s3.put_bucket_logging(Bucket=bucket_name, BucketLoggingStatus={
|
|
"LoggingEnabled": {
|
|
"TargetBucket": log_bucket,
|
|
"TargetPrefix": "{}/".format(bucket_name),
|
|
"TargetGrants": [
|
|
{
|
|
"Grantee": {
|
|
"ID": "SOMEIDSTRINGHERE9238748923734823917498237489237409123840983274",
|
|
"Type": "CanonicalUser"
|
|
},
|
|
"Permission": "NOTAREALPERM"
|
|
}
|
|
]
|
|
}
|
|
})
|
|
assert err.exception.response["Error"]["Code"] == "MalformedXML"
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_put_object_tagging():
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
bucket_name = 'mybucket'
|
|
key = 'key-with-tags'
|
|
s3.create_bucket(Bucket=bucket_name)
|
|
|
|
with assert_raises(ClientError) as err:
|
|
s3.put_object_tagging(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
Tagging={'TagSet': [
|
|
{'Key': 'item1', 'Value': 'foo'},
|
|
{'Key': 'item2', 'Value': 'bar'},
|
|
]}
|
|
)
|
|
|
|
e = err.exception
|
|
e.response['Error'].should.equal({
|
|
'Code': 'NoSuchKey',
|
|
'Message': 'The specified key does not exist.',
|
|
'RequestID': '7a62c49f-347e-4fc4-9331-6e8eEXAMPLE',
|
|
})
|
|
|
|
s3.put_object(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
Body='test'
|
|
)
|
|
|
|
resp = s3.put_object_tagging(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
Tagging={'TagSet': [
|
|
{'Key': 'item1', 'Value': 'foo'},
|
|
{'Key': 'item2', 'Value': 'bar'},
|
|
]}
|
|
)
|
|
|
|
resp['ResponseMetadata']['HTTPStatusCode'].should.equal(200)
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_put_object_tagging_with_single_tag():
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
bucket_name = 'mybucket'
|
|
key = 'key-with-tags'
|
|
s3.create_bucket(Bucket=bucket_name)
|
|
|
|
s3.put_object(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
Body='test'
|
|
)
|
|
|
|
resp = s3.put_object_tagging(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
Tagging={'TagSet': [
|
|
{'Key': 'item1', 'Value': 'foo'}
|
|
]}
|
|
)
|
|
|
|
resp['ResponseMetadata']['HTTPStatusCode'].should.equal(200)
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_get_object_tagging():
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
bucket_name = 'mybucket'
|
|
key = 'key-with-tags'
|
|
s3.create_bucket(Bucket=bucket_name)
|
|
|
|
s3.put_object(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
Body='test'
|
|
)
|
|
|
|
resp = s3.get_object_tagging(Bucket=bucket_name, Key=key)
|
|
resp['TagSet'].should.have.length_of(0)
|
|
|
|
resp = s3.put_object_tagging(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
Tagging={'TagSet': [
|
|
{'Key': 'item1', 'Value': 'foo'},
|
|
{'Key': 'item2', 'Value': 'bar'},
|
|
]}
|
|
)
|
|
resp = s3.get_object_tagging(Bucket=bucket_name, Key=key)
|
|
|
|
resp['TagSet'].should.have.length_of(2)
|
|
resp['TagSet'].should.contain({'Key': 'item1', 'Value': 'foo'})
|
|
resp['TagSet'].should.contain({'Key': 'item2', 'Value': 'bar'})
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_list_object_versions():
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
bucket_name = 'mybucket'
|
|
key = 'key-with-versions'
|
|
s3.create_bucket(Bucket=bucket_name)
|
|
s3.put_bucket_versioning(
|
|
Bucket=bucket_name,
|
|
VersioningConfiguration={
|
|
'Status': 'Enabled'
|
|
}
|
|
)
|
|
items = (six.b('v1'), six.b('v2'))
|
|
for body in items:
|
|
s3.put_object(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
Body=body
|
|
)
|
|
response = s3.list_object_versions(
|
|
Bucket=bucket_name
|
|
)
|
|
# Two object versions should be returned
|
|
len(response['Versions']).should.equal(2)
|
|
keys = set([item['Key'] for item in response['Versions']])
|
|
keys.should.equal({key})
|
|
# Test latest object version is returned
|
|
response = s3.get_object(Bucket=bucket_name, Key=key)
|
|
response['Body'].read().should.equal(items[-1])
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_bad_prefix_list_object_versions():
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
bucket_name = 'mybucket'
|
|
key = 'key-with-versions'
|
|
bad_prefix = 'key-that-does-not-exist'
|
|
s3.create_bucket(Bucket=bucket_name)
|
|
s3.put_bucket_versioning(
|
|
Bucket=bucket_name,
|
|
VersioningConfiguration={
|
|
'Status': 'Enabled'
|
|
}
|
|
)
|
|
items = (six.b('v1'), six.b('v2'))
|
|
for body in items:
|
|
s3.put_object(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
Body=body
|
|
)
|
|
response = s3.list_object_versions(
|
|
Bucket=bucket_name,
|
|
Prefix=bad_prefix,
|
|
)
|
|
response['ResponseMetadata']['HTTPStatusCode'].should.equal(200)
|
|
response.should_not.contain('Versions')
|
|
response.should_not.contain('DeleteMarkers')
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_delete_markers():
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
bucket_name = 'mybucket'
|
|
key = u'key-with-versions-and-unicode-ó'
|
|
s3.create_bucket(Bucket=bucket_name)
|
|
s3.put_bucket_versioning(
|
|
Bucket=bucket_name,
|
|
VersioningConfiguration={
|
|
'Status': 'Enabled'
|
|
}
|
|
)
|
|
items = (six.b('v1'), six.b('v2'))
|
|
for body in items:
|
|
s3.put_object(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
Body=body
|
|
)
|
|
|
|
s3.delete_objects(Bucket=bucket_name, Delete={'Objects': [{'Key': key}]})
|
|
|
|
with assert_raises(ClientError) as e:
|
|
s3.get_object(
|
|
Bucket=bucket_name,
|
|
Key=key
|
|
)
|
|
e.response['Error']['Code'].should.equal('404')
|
|
|
|
s3.delete_object(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
VersionId='2'
|
|
)
|
|
response = s3.get_object(
|
|
Bucket=bucket_name,
|
|
Key=key
|
|
)
|
|
response['Body'].read().should.equal(items[-1])
|
|
response = s3.list_object_versions(
|
|
Bucket=bucket_name
|
|
)
|
|
response['Versions'].should.have.length_of(2)
|
|
|
|
# We've asserted there is only 2 records so one is newest, one is oldest
|
|
latest = list(filter(lambda item: item['IsLatest'], response['Versions']))[0]
|
|
oldest = list(filter(lambda item: not item['IsLatest'], response['Versions']))[0]
|
|
|
|
# Double check ordering of version ID's
|
|
latest['VersionId'].should.equal('1')
|
|
oldest['VersionId'].should.equal('0')
|
|
|
|
# Double check the name is still unicode
|
|
latest['Key'].should.equal('key-with-versions-and-unicode-ó')
|
|
oldest['Key'].should.equal('key-with-versions-and-unicode-ó')
|
|
|
|
|
|
@mock_s3
|
|
def test_boto3_multiple_delete_markers():
|
|
s3 = boto3.client('s3', region_name='us-east-1')
|
|
bucket_name = 'mybucket'
|
|
key = u'key-with-versions-and-unicode-ó'
|
|
s3.create_bucket(Bucket=bucket_name)
|
|
s3.put_bucket_versioning(
|
|
Bucket=bucket_name,
|
|
VersioningConfiguration={
|
|
'Status': 'Enabled'
|
|
}
|
|
)
|
|
items = (six.b('v1'), six.b('v2'))
|
|
for body in items:
|
|
s3.put_object(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
Body=body
|
|
)
|
|
|
|
# Delete the object twice to add multiple delete markers
|
|
s3.delete_object(Bucket=bucket_name, Key=key)
|
|
s3.delete_object(Bucket=bucket_name, Key=key)
|
|
|
|
response = s3.list_object_versions(Bucket=bucket_name)
|
|
response['DeleteMarkers'].should.have.length_of(2)
|
|
|
|
with assert_raises(ClientError) as e:
|
|
s3.get_object(
|
|
Bucket=bucket_name,
|
|
Key=key
|
|
)
|
|
e.response['Error']['Code'].should.equal('404')
|
|
|
|
# Remove both delete markers to restore the object
|
|
s3.delete_object(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
VersionId='2'
|
|
)
|
|
s3.delete_object(
|
|
Bucket=bucket_name,
|
|
Key=key,
|
|
VersionId='3'
|
|
)
|
|
|
|
response = s3.get_object(
|
|
Bucket=bucket_name,
|
|
Key=key
|
|
)
|
|
response['Body'].read().should.equal(items[-1])
|
|
response = s3.list_object_versions(Bucket=bucket_name)
|
|
response['Versions'].should.have.length_of(2)
|
|
|
|
# We've asserted there is only 2 records so one is newest, one is oldest
|
|
latest = list(filter(lambda item: item['IsLatest'], response['Versions']))[0]
|
|
oldest = list(filter(lambda item: not item['IsLatest'], response['Versions']))[0]
|
|
|
|
# Double check ordering of version ID's
|
|
latest['VersionId'].should.equal('1')
|
|
oldest['VersionId'].should.equal('0')
|
|
|
|
# Double check the name is still unicode
|
|
latest['Key'].should.equal('key-with-versions-and-unicode-ó')
|
|
oldest['Key'].should.equal('key-with-versions-and-unicode-ó')
|
|
|
|
@mock_s3
|
|
def test_get_stream_gzipped():
|
|
payload = b"this is some stuff here"
|
|
|
|
s3_client = boto3.client("s3", region_name='us-east-1')
|
|
s3_client.create_bucket(Bucket='moto-tests')
|
|
buffer_ = BytesIO()
|
|
with GzipFile(fileobj=buffer_, mode='w') as f:
|
|
f.write(payload)
|
|
payload_gz = buffer_.getvalue()
|
|
|
|
s3_client.put_object(
|
|
Bucket='moto-tests',
|
|
Key='keyname',
|
|
Body=payload_gz,
|
|
ContentEncoding='gzip',
|
|
)
|
|
|
|
obj = s3_client.get_object(
|
|
Bucket='moto-tests',
|
|
Key='keyname',
|
|
)
|
|
res = zlib.decompress(obj['Body'].read(), 16 + zlib.MAX_WBITS)
|
|
assert res == payload
|
|
|
|
|
|
TEST_XML = """\
|
|
<?xml version="1.0" encoding="UTF-8"?>
|
|
<ns0:WebsiteConfiguration xmlns:ns0="http://s3.amazonaws.com/doc/2006-03-01/">
|
|
<ns0:IndexDocument>
|
|
<ns0:Suffix>index.html</ns0:Suffix>
|
|
</ns0:IndexDocument>
|
|
<ns0:RoutingRules>
|
|
<ns0:RoutingRule>
|
|
<ns0:Condition>
|
|
<ns0:KeyPrefixEquals>test/testing</ns0:KeyPrefixEquals>
|
|
</ns0:Condition>
|
|
<ns0:Redirect>
|
|
<ns0:ReplaceKeyWith>test.txt</ns0:ReplaceKeyWith>
|
|
</ns0:Redirect>
|
|
</ns0:RoutingRule>
|
|
</ns0:RoutingRules>
|
|
</ns0:WebsiteConfiguration>
|
|
"""
|