Merge branch 'master' of https://github.com/spulec/moto
This commit is contained in:
commit
bd9e7deb95
14
README.md
14
README.md
@ -200,16 +200,24 @@ In general, Moto doesn't rely on anything specific to Boto. It only mocks AWS en
|
|||||||
|
|
||||||
## Stand-alone Server Mode
|
## Stand-alone Server Mode
|
||||||
|
|
||||||
Moto also comes with a stand-alone server mode. This allows you to utilize the backend structure of Moto even if you don't use Python.
|
Moto also has a stand-alone server mode. This allows you to utilize
|
||||||
|
the backend structure of Moto even if you don't use Python.
|
||||||
|
|
||||||
To run a service:
|
It uses flask, which isn't a default dependency. You can install the
|
||||||
|
server 'extra' package with:
|
||||||
|
|
||||||
|
```python
|
||||||
|
pip install moto[server]
|
||||||
|
```
|
||||||
|
|
||||||
|
You can then start it running a service:
|
||||||
|
|
||||||
```console
|
```console
|
||||||
$ moto_server ec2
|
$ moto_server ec2
|
||||||
* Running on http://127.0.0.1:5000/
|
* Running on http://127.0.0.1:5000/
|
||||||
```
|
```
|
||||||
|
|
||||||
You can also pass the port as the second argument:
|
You can also pass the port:
|
||||||
|
|
||||||
```console
|
```console
|
||||||
$ moto_server ec2 -p3000
|
$ moto_server ec2 -p3000
|
||||||
|
@ -3,7 +3,15 @@ from __future__ import unicode_literals
|
|||||||
import base64
|
import base64
|
||||||
import datetime
|
import datetime
|
||||||
import hashlib
|
import hashlib
|
||||||
|
import io
|
||||||
import json
|
import json
|
||||||
|
import sys
|
||||||
|
import zipfile
|
||||||
|
|
||||||
|
try:
|
||||||
|
from StringIO import StringIO
|
||||||
|
except:
|
||||||
|
from io import StringIO
|
||||||
|
|
||||||
import boto.awslambda
|
import boto.awslambda
|
||||||
from moto.core import BaseBackend
|
from moto.core import BaseBackend
|
||||||
@ -34,9 +42,18 @@ class LambdaFunction(object):
|
|||||||
self.version = '$LATEST'
|
self.version = '$LATEST'
|
||||||
self.last_modified = datetime.datetime.utcnow().strftime('%Y-%m-%d %H:%M:%S')
|
self.last_modified = datetime.datetime.utcnow().strftime('%Y-%m-%d %H:%M:%S')
|
||||||
if 'ZipFile' in self.code:
|
if 'ZipFile' in self.code:
|
||||||
code = base64.b64decode(self.code['ZipFile'])
|
# more hackery to handle unicode/bytes/str in python3 and python2 - argh!
|
||||||
self.code_size = len(code)
|
try:
|
||||||
self.code_sha_256 = hashlib.sha256(code).hexdigest()
|
to_unzip_code = base64.b64decode(bytes(self.code['ZipFile'], 'utf-8'))
|
||||||
|
except Exception:
|
||||||
|
to_unzip_code = base64.b64decode(self.code['ZipFile'])
|
||||||
|
|
||||||
|
zbuffer = io.BytesIO()
|
||||||
|
zbuffer.write(to_unzip_code)
|
||||||
|
zip_file = zipfile.ZipFile(zbuffer, 'r', zipfile.ZIP_DEFLATED)
|
||||||
|
self.code = zip_file.read("".join(zip_file.namelist()))
|
||||||
|
self.code_size = len(to_unzip_code)
|
||||||
|
self.code_sha_256 = hashlib.sha256(to_unzip_code).hexdigest()
|
||||||
else:
|
else:
|
||||||
# validate s3 bucket
|
# validate s3 bucket
|
||||||
try:
|
try:
|
||||||
@ -93,15 +110,56 @@ class LambdaFunction(object):
|
|||||||
"Configuration": self.get_configuration(),
|
"Configuration": self.get_configuration(),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
def convert(self, s):
|
||||||
|
try:
|
||||||
|
return str(s, encoding='utf8')
|
||||||
|
except:
|
||||||
|
return s
|
||||||
|
|
||||||
|
def is_json(self, test_str):
|
||||||
|
try:
|
||||||
|
response = json.loads(test_str)
|
||||||
|
except:
|
||||||
|
response = test_str
|
||||||
|
return response
|
||||||
|
|
||||||
|
def _invoke_lambda(self, code, event={}, context={}):
|
||||||
|
# TO DO: context not yet implemented
|
||||||
|
try:
|
||||||
|
mycode = "\n".join(['import json',
|
||||||
|
self.convert(self.code),
|
||||||
|
self.convert('print(lambda_handler(%s, %s))' % (self.is_json(self.convert(event)), context))])
|
||||||
|
#print("moto_lambda_debug: ", mycode)
|
||||||
|
except Exception as ex:
|
||||||
|
print("Exception %s", ex)
|
||||||
|
|
||||||
|
try:
|
||||||
|
codeOut = StringIO()
|
||||||
|
codeErr = StringIO()
|
||||||
|
sys.stdout = codeOut
|
||||||
|
sys.stderr = codeErr
|
||||||
|
exec(mycode)
|
||||||
|
exec_err = codeErr.getvalue()
|
||||||
|
exec_out = codeOut.getvalue()
|
||||||
|
result = "\n".join([exec_out, self.convert(exec_err)])
|
||||||
|
except Exception as ex:
|
||||||
|
result = '%s\n\n\nException %s' % (mycode, ex)
|
||||||
|
finally:
|
||||||
|
codeErr.close()
|
||||||
|
codeOut.close()
|
||||||
|
sys.stdout = sys.__stdout__
|
||||||
|
sys.stderr = sys.__stderr__
|
||||||
|
return self.convert(result)
|
||||||
|
|
||||||
def invoke(self, request, headers):
|
def invoke(self, request, headers):
|
||||||
payload = dict()
|
payload = dict()
|
||||||
|
|
||||||
# Get the invocation type:
|
# Get the invocation type:
|
||||||
|
r = self._invoke_lambda(code=self.code, event=request.body)
|
||||||
if request.headers.get("x-amz-invocation-type") == "RequestResponse":
|
if request.headers.get("x-amz-invocation-type") == "RequestResponse":
|
||||||
encoded = base64.b64encode("Some log file output...".encode('utf-8'))
|
encoded = base64.b64encode(r.encode('utf-8'))
|
||||||
headers["x-amz-log-result"] = encoded.decode('utf-8')
|
headers["x-amz-log-result"] = encoded.decode('utf-8')
|
||||||
|
payload['result'] = headers["x-amz-log-result"]
|
||||||
payload["result"] = "Good"
|
|
||||||
|
|
||||||
return json.dumps(payload, indent=4)
|
return json.dumps(payload, indent=4)
|
||||||
|
|
||||||
@ -154,3 +212,7 @@ class LambdaBackend(BaseBackend):
|
|||||||
lambda_backends = {}
|
lambda_backends = {}
|
||||||
for region in boto.awslambda.regions():
|
for region in boto.awslambda.regions():
|
||||||
lambda_backends[region.name] = LambdaBackend()
|
lambda_backends[region.name] = LambdaBackend()
|
||||||
|
|
||||||
|
# Handle us forgotten regions, unless Lambda truly only runs out of US and EU?????
|
||||||
|
for region in ['ap-southeast-2']:
|
||||||
|
lambda_backends[region] = LambdaBackend()
|
||||||
|
@ -43,7 +43,7 @@ class LambdaResponse(BaseResponse):
|
|||||||
if lambda_backend.has_function(function_name):
|
if lambda_backend.has_function(function_name):
|
||||||
fn = lambda_backend.get_function(function_name)
|
fn = lambda_backend.get_function(function_name)
|
||||||
payload = fn.invoke(request, headers)
|
payload = fn.invoke(request, headers)
|
||||||
return 200, headers, payload
|
return 202, headers, payload
|
||||||
else:
|
else:
|
||||||
return 404, headers, "{}"
|
return 404, headers, "{}"
|
||||||
|
|
||||||
|
@ -275,9 +275,14 @@ class Table(object):
|
|||||||
raise ValueError("The conditional request failed")
|
raise ValueError("The conditional request failed")
|
||||||
elif key not in current_attr:
|
elif key not in current_attr:
|
||||||
raise ValueError("The conditional request failed")
|
raise ValueError("The conditional request failed")
|
||||||
elif DynamoType(val['Value']).value != current_attr[key].value:
|
elif 'Value' in val and DynamoType(val['Value']).value != current_attr[key].value:
|
||||||
raise ValueError("The conditional request failed")
|
raise ValueError("The conditional request failed")
|
||||||
|
elif 'ComparisonOperator' in val:
|
||||||
|
comparison_func = get_comparison_func(val['ComparisonOperator'])
|
||||||
|
dynamo_types = [DynamoType(ele) for ele in val["AttributeValueList"]]
|
||||||
|
for t in dynamo_types:
|
||||||
|
if not comparison_func(current_attr[key].value, t.value):
|
||||||
|
raise ValueError('The conditional request failed')
|
||||||
if range_value:
|
if range_value:
|
||||||
self.items[hash_value][range_value] = item
|
self.items[hash_value][range_value] = item
|
||||||
else:
|
else:
|
||||||
|
@ -179,14 +179,14 @@ class DynamoHandler(BaseResponse):
|
|||||||
item = dynamodb_backend2.delete_item(table_name, keys)
|
item = dynamodb_backend2.delete_item(table_name, keys)
|
||||||
|
|
||||||
response = {
|
response = {
|
||||||
"Responses": {
|
"ConsumedCapacity": [
|
||||||
"Thread": {
|
{
|
||||||
"ConsumedCapacityUnits": 1.0
|
'TableName': table_name,
|
||||||
},
|
'CapacityUnits': 1.0,
|
||||||
"Reply": {
|
'Table': {'CapacityUnits': 1.0}
|
||||||
"ConsumedCapacityUnits": 1.0
|
} for table_name, table_requests in table_batches.items()
|
||||||
}
|
],
|
||||||
},
|
"ItemCollectionMetrics": {},
|
||||||
"UnprocessedItems": {}
|
"UnprocessedItems": {}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -632,6 +632,8 @@ class InstanceBackend(object):
|
|||||||
|
|
||||||
def terminate_instances(self, instance_ids):
|
def terminate_instances(self, instance_ids):
|
||||||
terminated_instances = []
|
terminated_instances = []
|
||||||
|
if not instance_ids:
|
||||||
|
raise EC2ClientError("InvalidParameterCombination", "No instances specified")
|
||||||
for instance in self.get_multi_instances_by_id(instance_ids):
|
for instance in self.get_multi_instances_by_id(instance_ids):
|
||||||
instance.terminate()
|
instance.terminate()
|
||||||
terminated_instances.append(instance)
|
terminated_instances.append(instance)
|
||||||
|
@ -176,17 +176,23 @@ class FirehoseRecord(object):
|
|||||||
class DeliveryStream(object):
|
class DeliveryStream(object):
|
||||||
def __init__(self, stream_name, **stream_kwargs):
|
def __init__(self, stream_name, **stream_kwargs):
|
||||||
self.name = stream_name
|
self.name = stream_name
|
||||||
self.redshift_username = stream_kwargs['redshift_username']
|
self.redshift_username = stream_kwargs.get('redshift_username')
|
||||||
self.redshift_password = stream_kwargs['redshift_password']
|
self.redshift_password = stream_kwargs.get('redshift_password')
|
||||||
self.redshift_jdbc_url = stream_kwargs['redshift_jdbc_url']
|
self.redshift_jdbc_url = stream_kwargs.get('redshift_jdbc_url')
|
||||||
self.redshift_role_arn = stream_kwargs['redshift_role_arn']
|
self.redshift_role_arn = stream_kwargs.get('redshift_role_arn')
|
||||||
self.redshift_copy_command = stream_kwargs['redshift_copy_command']
|
self.redshift_copy_command = stream_kwargs.get('redshift_copy_command')
|
||||||
|
|
||||||
self.redshift_s3_role_arn = stream_kwargs['redshift_s3_role_arn']
|
self.s3_role_arn = stream_kwargs.get('s3_role_arn')
|
||||||
self.redshift_s3_bucket_arn = stream_kwargs['redshift_s3_bucket_arn']
|
self.s3_bucket_arn = stream_kwargs.get('s3_bucket_arn')
|
||||||
self.redshift_s3_prefix = stream_kwargs['redshift_s3_prefix']
|
self.s3_prefix = stream_kwargs.get('s3_prefix')
|
||||||
|
self.s3_compression_format = stream_kwargs.get('s3_compression_format', 'UNCOMPRESSED')
|
||||||
|
self.s3_buffering_hings = stream_kwargs.get('s3_buffering_hings')
|
||||||
|
|
||||||
|
self.redshift_s3_role_arn = stream_kwargs.get('redshift_s3_role_arn')
|
||||||
|
self.redshift_s3_bucket_arn = stream_kwargs.get('redshift_s3_bucket_arn')
|
||||||
|
self.redshift_s3_prefix = stream_kwargs.get('redshift_s3_prefix')
|
||||||
self.redshift_s3_compression_format = stream_kwargs.get('redshift_s3_compression_format', 'UNCOMPRESSED')
|
self.redshift_s3_compression_format = stream_kwargs.get('redshift_s3_compression_format', 'UNCOMPRESSED')
|
||||||
self.redshift_s3_buffering_hings = stream_kwargs['redshift_s3_buffering_hings']
|
self.redshift_s3_buffering_hings = stream_kwargs.get('redshift_s3_buffering_hings')
|
||||||
|
|
||||||
self.records = []
|
self.records = []
|
||||||
self.status = 'ACTIVE'
|
self.status = 'ACTIVE'
|
||||||
@ -197,6 +203,38 @@ class DeliveryStream(object):
|
|||||||
def arn(self):
|
def arn(self):
|
||||||
return 'arn:aws:firehose:us-east-1:123456789012:deliverystream/{0}'.format(self.name)
|
return 'arn:aws:firehose:us-east-1:123456789012:deliverystream/{0}'.format(self.name)
|
||||||
|
|
||||||
|
def destinations_to_dict(self):
|
||||||
|
if self.s3_role_arn:
|
||||||
|
return [{
|
||||||
|
'DestinationId': 'string',
|
||||||
|
'S3DestinationDescription': {
|
||||||
|
'RoleARN': self.s3_role_arn,
|
||||||
|
'BucketARN': self.s3_bucket_arn,
|
||||||
|
'Prefix': self.s3_prefix,
|
||||||
|
'BufferingHints': self.s3_buffering_hings,
|
||||||
|
'CompressionFormat': self.s3_compression_format,
|
||||||
|
}
|
||||||
|
}]
|
||||||
|
else:
|
||||||
|
return [{
|
||||||
|
"DestinationId": "string",
|
||||||
|
"RedshiftDestinationDescription": {
|
||||||
|
"ClusterJDBCURL": self.redshift_jdbc_url,
|
||||||
|
"CopyCommand": self.redshift_copy_command,
|
||||||
|
"RoleARN": self.redshift_role_arn,
|
||||||
|
"S3DestinationDescription": {
|
||||||
|
"BucketARN": self.redshift_s3_bucket_arn,
|
||||||
|
"BufferingHints": self.redshift_s3_buffering_hings,
|
||||||
|
"CompressionFormat": self.redshift_s3_compression_format,
|
||||||
|
"Prefix": self.redshift_s3_prefix,
|
||||||
|
"RoleARN": self.redshift_s3_role_arn
|
||||||
|
},
|
||||||
|
"Username": self.redshift_username,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
def to_dict(self):
|
def to_dict(self):
|
||||||
return {
|
return {
|
||||||
"DeliveryStreamDescription": {
|
"DeliveryStreamDescription": {
|
||||||
@ -204,24 +242,7 @@ class DeliveryStream(object):
|
|||||||
"DeliveryStreamARN": self.arn,
|
"DeliveryStreamARN": self.arn,
|
||||||
"DeliveryStreamName": self.name,
|
"DeliveryStreamName": self.name,
|
||||||
"DeliveryStreamStatus": self.status,
|
"DeliveryStreamStatus": self.status,
|
||||||
"Destinations": [
|
"Destinations": self.destinations_to_dict(),
|
||||||
{
|
|
||||||
"DestinationId": "string",
|
|
||||||
"RedshiftDestinationDescription": {
|
|
||||||
"ClusterJDBCURL": self.redshift_jdbc_url,
|
|
||||||
"CopyCommand": self.redshift_copy_command,
|
|
||||||
"RoleARN": self.redshift_role_arn,
|
|
||||||
"S3DestinationDescription": {
|
|
||||||
"BucketARN": self.redshift_s3_bucket_arn,
|
|
||||||
"BufferingHints": self.redshift_s3_buffering_hings,
|
|
||||||
"CompressionFormat": self.redshift_s3_compression_format,
|
|
||||||
"Prefix": self.redshift_s3_prefix,
|
|
||||||
"RoleARN": self.redshift_s3_role_arn
|
|
||||||
},
|
|
||||||
"Username": self.redshift_username,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
],
|
|
||||||
"HasMoreDestinations": False,
|
"HasMoreDestinations": False,
|
||||||
"LastUpdateTimestamp": time.mktime(self.last_updated.timetuple()),
|
"LastUpdateTimestamp": time.mktime(self.last_updated.timetuple()),
|
||||||
"VersionId": "string",
|
"VersionId": "string",
|
||||||
|
@ -139,6 +139,16 @@ class KinesisResponse(BaseResponse):
|
|||||||
'redshift_s3_compression_format': redshift_s3_config.get('CompressionFormat'),
|
'redshift_s3_compression_format': redshift_s3_config.get('CompressionFormat'),
|
||||||
'redshift_s3_buffering_hings': redshift_s3_config['BufferingHints'],
|
'redshift_s3_buffering_hings': redshift_s3_config['BufferingHints'],
|
||||||
}
|
}
|
||||||
|
else:
|
||||||
|
# S3 Config
|
||||||
|
s3_config = self.parameters['S3DestinationConfiguration']
|
||||||
|
stream_kwargs = {
|
||||||
|
's3_role_arn': s3_config['RoleARN'],
|
||||||
|
's3_bucket_arn': s3_config['BucketARN'],
|
||||||
|
's3_prefix': s3_config['Prefix'],
|
||||||
|
's3_compression_format': s3_config.get('CompressionFormat'),
|
||||||
|
's3_buffering_hings': s3_config['BufferingHints'],
|
||||||
|
}
|
||||||
stream = self.kinesis_backend.create_delivery_stream(stream_name, **stream_kwargs)
|
stream = self.kinesis_backend.create_delivery_stream(stream_name, **stream_kwargs)
|
||||||
return json.dumps({
|
return json.dumps({
|
||||||
'DeliveryStreamARN': stream.arn
|
'DeliveryStreamARN': stream.arn
|
||||||
|
@ -77,11 +77,19 @@ class KmsBackend(BaseBackend):
|
|||||||
return self.keys.pop(key_id)
|
return self.keys.pop(key_id)
|
||||||
|
|
||||||
def describe_key(self, key_id):
|
def describe_key(self, key_id):
|
||||||
return self.keys[key_id]
|
# allow the different methods (alias, ARN :key/, keyId, ARN alias) to describe key not just KeyId
|
||||||
|
key_id = self.get_key_id(key_id)
|
||||||
|
if r'alias/' in str(key_id).lower():
|
||||||
|
key_id = self.get_key_id_from_alias(key_id.split('alias/')[1])
|
||||||
|
return self.keys[self.get_key_id(key_id)]
|
||||||
|
|
||||||
def list_keys(self):
|
def list_keys(self):
|
||||||
return self.keys.values()
|
return self.keys.values()
|
||||||
|
|
||||||
|
def get_key_id(self, key_id):
|
||||||
|
# Allow use of ARN as well as pure KeyId
|
||||||
|
return str(key_id).split(r':key/')[1] if r':key/' in str(key_id).lower() else key_id
|
||||||
|
|
||||||
def alias_exists(self, alias_name):
|
def alias_exists(self, alias_name):
|
||||||
for aliases in self.key_to_aliases.values():
|
for aliases in self.key_to_aliases.values():
|
||||||
if alias_name in aliases:
|
if alias_name in aliases:
|
||||||
@ -99,21 +107,26 @@ class KmsBackend(BaseBackend):
|
|||||||
def get_all_aliases(self):
|
def get_all_aliases(self):
|
||||||
return self.key_to_aliases
|
return self.key_to_aliases
|
||||||
|
|
||||||
|
def get_key_id_from_alias(self, alias_name):
|
||||||
|
for key_id, aliases in dict(self.key_to_aliases).items():
|
||||||
|
if alias_name in ",".join(aliases):
|
||||||
|
return key_id
|
||||||
|
return None
|
||||||
|
|
||||||
def enable_key_rotation(self, key_id):
|
def enable_key_rotation(self, key_id):
|
||||||
self.keys[key_id].key_rotation_status = True
|
self.keys[self.get_key_id(key_id)].key_rotation_status = True
|
||||||
|
|
||||||
def disable_key_rotation(self, key_id):
|
def disable_key_rotation(self, key_id):
|
||||||
self.keys[key_id].key_rotation_status = False
|
self.keys[self.get_key_id(key_id)].key_rotation_status = False
|
||||||
|
|
||||||
def get_key_rotation_status(self, key_id):
|
def get_key_rotation_status(self, key_id):
|
||||||
return self.keys[key_id].key_rotation_status
|
return self.keys[self.get_key_id(key_id)].key_rotation_status
|
||||||
|
|
||||||
def put_key_policy(self, key_id, policy):
|
def put_key_policy(self, key_id, policy):
|
||||||
self.keys[key_id].policy = policy
|
self.keys[self.get_key_id(key_id)].policy = policy
|
||||||
|
|
||||||
def get_key_policy(self, key_id):
|
def get_key_policy(self, key_id):
|
||||||
return self.keys[key_id].policy
|
return self.keys[self.get_key_id(key_id)].policy
|
||||||
|
|
||||||
|
|
||||||
kms_backends = {}
|
kms_backends = {}
|
||||||
for region in boto.kms.regions():
|
for region in boto.kms.regions():
|
||||||
|
@ -37,7 +37,7 @@ class KmsResponse(BaseResponse):
|
|||||||
def describe_key(self):
|
def describe_key(self):
|
||||||
key_id = self.parameters.get('KeyId')
|
key_id = self.parameters.get('KeyId')
|
||||||
try:
|
try:
|
||||||
key = self.kms_backend.describe_key(key_id)
|
key = self.kms_backend.describe_key(self.kms_backend.get_key_id(key_id))
|
||||||
except KeyError:
|
except KeyError:
|
||||||
headers = dict(self.headers)
|
headers = dict(self.headers)
|
||||||
headers['status'] = 404
|
headers['status'] = 404
|
||||||
@ -140,7 +140,7 @@ class KmsResponse(BaseResponse):
|
|||||||
|
|
||||||
def enable_key_rotation(self):
|
def enable_key_rotation(self):
|
||||||
key_id = self.parameters.get('KeyId')
|
key_id = self.parameters.get('KeyId')
|
||||||
_assert_valid_key_id(key_id)
|
_assert_valid_key_id(self.kms_backend.get_key_id(key_id))
|
||||||
try:
|
try:
|
||||||
self.kms_backend.enable_key_rotation(key_id)
|
self.kms_backend.enable_key_rotation(key_id)
|
||||||
except KeyError:
|
except KeyError:
|
||||||
@ -152,7 +152,7 @@ class KmsResponse(BaseResponse):
|
|||||||
|
|
||||||
def disable_key_rotation(self):
|
def disable_key_rotation(self):
|
||||||
key_id = self.parameters.get('KeyId')
|
key_id = self.parameters.get('KeyId')
|
||||||
_assert_valid_key_id(key_id)
|
_assert_valid_key_id(self.kms_backend.get_key_id(key_id))
|
||||||
try:
|
try:
|
||||||
self.kms_backend.disable_key_rotation(key_id)
|
self.kms_backend.disable_key_rotation(key_id)
|
||||||
except KeyError:
|
except KeyError:
|
||||||
@ -163,7 +163,7 @@ class KmsResponse(BaseResponse):
|
|||||||
|
|
||||||
def get_key_rotation_status(self):
|
def get_key_rotation_status(self):
|
||||||
key_id = self.parameters.get('KeyId')
|
key_id = self.parameters.get('KeyId')
|
||||||
_assert_valid_key_id(key_id)
|
_assert_valid_key_id(self.kms_backend.get_key_id(key_id))
|
||||||
try:
|
try:
|
||||||
rotation_enabled = self.kms_backend.get_key_rotation_status(key_id)
|
rotation_enabled = self.kms_backend.get_key_rotation_status(key_id)
|
||||||
except KeyError:
|
except KeyError:
|
||||||
@ -176,7 +176,7 @@ class KmsResponse(BaseResponse):
|
|||||||
key_id = self.parameters.get('KeyId')
|
key_id = self.parameters.get('KeyId')
|
||||||
policy_name = self.parameters.get('PolicyName')
|
policy_name = self.parameters.get('PolicyName')
|
||||||
policy = self.parameters.get('Policy')
|
policy = self.parameters.get('Policy')
|
||||||
_assert_valid_key_id(key_id)
|
_assert_valid_key_id(self.kms_backend.get_key_id(key_id))
|
||||||
_assert_default_policy(policy_name)
|
_assert_default_policy(policy_name)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
@ -191,7 +191,7 @@ class KmsResponse(BaseResponse):
|
|||||||
def get_key_policy(self):
|
def get_key_policy(self):
|
||||||
key_id = self.parameters.get('KeyId')
|
key_id = self.parameters.get('KeyId')
|
||||||
policy_name = self.parameters.get('PolicyName')
|
policy_name = self.parameters.get('PolicyName')
|
||||||
_assert_valid_key_id(key_id)
|
_assert_valid_key_id(self.kms_backend.get_key_id(key_id))
|
||||||
_assert_default_policy(policy_name)
|
_assert_default_policy(policy_name)
|
||||||
|
|
||||||
try:
|
try:
|
||||||
@ -203,7 +203,7 @@ class KmsResponse(BaseResponse):
|
|||||||
|
|
||||||
def list_key_policies(self):
|
def list_key_policies(self):
|
||||||
key_id = self.parameters.get('KeyId')
|
key_id = self.parameters.get('KeyId')
|
||||||
_assert_valid_key_id(key_id)
|
_assert_valid_key_id(self.kms_backend.get_key_id(key_id))
|
||||||
try:
|
try:
|
||||||
self.kms_backend.describe_key(key_id)
|
self.kms_backend.describe_key(key_id)
|
||||||
except KeyError:
|
except KeyError:
|
||||||
|
@ -238,18 +238,19 @@ class Route53Backend(BaseBackend):
|
|||||||
|
|
||||||
def change_tags_for_resource(self, resource_id, tags):
|
def change_tags_for_resource(self, resource_id, tags):
|
||||||
if 'Tag' in tags:
|
if 'Tag' in tags:
|
||||||
for key, tag in tags.items():
|
if isinstance(tags['Tag'], list):
|
||||||
for t in tag:
|
for tag in tags['Tag']:
|
||||||
self.resource_tags[resource_id][t['Key']] = t['Value']
|
self.resource_tags[resource_id][tag['Key']] = tag['Value']
|
||||||
|
else:
|
||||||
|
key, value = (tags['Tag']['Key'], tags['Tag']['Value'])
|
||||||
|
self.resource_tags[resource_id][key] = value
|
||||||
else:
|
else:
|
||||||
for _, keys in tags.items():
|
if 'Key' in tags:
|
||||||
if isinstance(keys, list):
|
if isinstance(tags['Key'], list):
|
||||||
for key in keys:
|
for key in tags['Key']:
|
||||||
del(self.resource_tags[resource_id][key])
|
del(self.resource_tags[resource_id][key])
|
||||||
else:
|
else:
|
||||||
del(self.resource_tags[resource_id][keys])
|
del(self.resource_tags[resource_id][tags['Key']])
|
||||||
|
|
||||||
|
|
||||||
def list_tags_for_resource(self, resource_id):
|
def list_tags_for_resource(self, resource_id):
|
||||||
if resource_id in self.resource_tags:
|
if resource_id in self.resource_tags:
|
||||||
|
5
setup.py
5
setup.py
@ -5,19 +5,18 @@ from setuptools import setup, find_packages
|
|||||||
install_requires = [
|
install_requires = [
|
||||||
"Jinja2",
|
"Jinja2",
|
||||||
"boto>=2.36.0",
|
"boto>=2.36.0",
|
||||||
"flask",
|
|
||||||
"httpretty==0.8.10",
|
"httpretty==0.8.10",
|
||||||
"requests",
|
"requests",
|
||||||
"xmltodict",
|
"xmltodict",
|
||||||
"six",
|
"six",
|
||||||
"werkzeug",
|
"werkzeug",
|
||||||
"sure",
|
|
||||||
"freezegun"
|
|
||||||
]
|
]
|
||||||
|
|
||||||
extras_require = {
|
extras_require = {
|
||||||
# No builtin OrderedDict before 2.7
|
# No builtin OrderedDict before 2.7
|
||||||
':python_version=="2.6"': ['ordereddict'],
|
':python_version=="2.6"': ['ordereddict'],
|
||||||
|
|
||||||
|
'server': ['flask'],
|
||||||
}
|
}
|
||||||
|
|
||||||
setup(
|
setup(
|
||||||
|
@ -1,51 +1,66 @@
|
|||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
import base64
|
||||||
import botocore.client
|
import botocore.client
|
||||||
import boto3
|
import boto3
|
||||||
import hashlib
|
import hashlib
|
||||||
import io
|
import io
|
||||||
|
import json
|
||||||
import zipfile
|
import zipfile
|
||||||
import sure # noqa
|
import sure # noqa
|
||||||
|
|
||||||
from freezegun import freeze_time
|
from freezegun import freeze_time
|
||||||
from moto import mock_lambda, mock_s3
|
from moto import mock_lambda, mock_s3, mock_ec2
|
||||||
|
|
||||||
|
|
||||||
def get_test_zip_file():
|
def _process_lamda(pfunc):
|
||||||
zip_output = io.BytesIO()
|
zip_output = io.BytesIO()
|
||||||
zip_file = zipfile.ZipFile(zip_output, 'w')
|
zip_file = zipfile.ZipFile(zip_output, 'w', zipfile.ZIP_DEFLATED)
|
||||||
zip_file.writestr('lambda_function.py', b'''\
|
zip_file.writestr('lambda_function.zip', pfunc)
|
||||||
def handler(event, context):
|
|
||||||
return "hello world"
|
|
||||||
''')
|
|
||||||
zip_file.close()
|
zip_file.close()
|
||||||
zip_output.seek(0)
|
zip_output.seek(0)
|
||||||
return zip_output.read()
|
return zip_output.read()
|
||||||
|
|
||||||
|
|
||||||
@mock_lambda
|
def get_test_zip_file1():
|
||||||
def test_list_functions():
|
pfunc = """
|
||||||
conn = boto3.client('lambda', 'us-west-2')
|
def lambda_handler(event, context):
|
||||||
|
return (event, context)
|
||||||
|
"""
|
||||||
|
return _process_lamda(pfunc)
|
||||||
|
|
||||||
result = conn.list_functions()
|
|
||||||
|
|
||||||
result['Functions'].should.have.length_of(0)
|
def get_test_zip_file2():
|
||||||
|
pfunc = """
|
||||||
|
def lambda_handler(event, context):
|
||||||
|
volume_id = event.get('volume_id')
|
||||||
|
print('get volume details for %s' % volume_id)
|
||||||
|
import boto3
|
||||||
|
ec2 = boto3.resource('ec2', region_name='us-west-2')
|
||||||
|
vol = ec2.Volume(volume_id)
|
||||||
|
print('Volume - %s state=%s, size=%s' % (volume_id, vol.state, vol.size))
|
||||||
|
"""
|
||||||
|
return _process_lamda(pfunc)
|
||||||
|
|
||||||
|
|
||||||
@mock_lambda
|
@mock_lambda
|
||||||
@mock_s3
|
@mock_s3
|
||||||
@freeze_time('2015-01-01 00:00:00')
|
def test_list_functions():
|
||||||
def test_invoke_function():
|
|
||||||
conn = boto3.client('lambda', 'us-west-2')
|
conn = boto3.client('lambda', 'us-west-2')
|
||||||
|
result = conn.list_functions()
|
||||||
|
result['Functions'].should.have.length_of(0)
|
||||||
|
|
||||||
zip_content = get_test_zip_file()
|
@mock_lambda
|
||||||
|
@freeze_time('2015-01-01 00:00:00')
|
||||||
|
def test_invoke_event_function():
|
||||||
|
conn = boto3.client('lambda', 'us-west-2')
|
||||||
conn.create_function(
|
conn.create_function(
|
||||||
FunctionName='testFunction',
|
FunctionName='testFunction',
|
||||||
Runtime='python2.7',
|
Runtime='python2.7',
|
||||||
Role='test-iam-role',
|
Role='test-iam-role',
|
||||||
Handler='lambda_function.handler',
|
Handler='lambda_function.handler',
|
||||||
Code={
|
Code={
|
||||||
'ZipFile': zip_content,
|
'ZipFile': get_test_zip_file1(),
|
||||||
},
|
},
|
||||||
Description='test lambda function',
|
Description='test lambda function',
|
||||||
Timeout=3,
|
Timeout=3,
|
||||||
@ -53,8 +68,8 @@ def test_invoke_function():
|
|||||||
Publish=True,
|
Publish=True,
|
||||||
)
|
)
|
||||||
|
|
||||||
success_result = conn.invoke(FunctionName='testFunction', InvocationType='Event', Payload='{}')
|
success_result = conn.invoke(FunctionName='testFunction', InvocationType='Event', Payload=json.dumps({'msg': 'Mostly Harmless'}))
|
||||||
success_result["StatusCode"].should.equal(200)
|
success_result["StatusCode"].should.equal(202)
|
||||||
|
|
||||||
conn.invoke.when.called_with(
|
conn.invoke.when.called_with(
|
||||||
FunctionName='notAFunction',
|
FunctionName='notAFunction',
|
||||||
@ -62,11 +77,63 @@ def test_invoke_function():
|
|||||||
Payload='{}'
|
Payload='{}'
|
||||||
).should.throw(botocore.client.ClientError)
|
).should.throw(botocore.client.ClientError)
|
||||||
|
|
||||||
success_result = conn.invoke(FunctionName='testFunction', InvocationType='RequestResponse', Payload='{}')
|
|
||||||
success_result["StatusCode"].should.equal(200)
|
@mock_lambda
|
||||||
|
@freeze_time('2015-01-01 00:00:00')
|
||||||
|
def test_invoke_requestresponse_function():
|
||||||
|
conn = boto3.client('lambda', 'us-west-2')
|
||||||
|
conn.create_function(
|
||||||
|
FunctionName='testFunction',
|
||||||
|
Runtime='python2.7',
|
||||||
|
Role='test-iam-role',
|
||||||
|
Handler='lambda_function.handler',
|
||||||
|
Code={
|
||||||
|
'ZipFile': get_test_zip_file1(),
|
||||||
|
},
|
||||||
|
Description='test lambda function',
|
||||||
|
Timeout=3,
|
||||||
|
MemorySize=128,
|
||||||
|
Publish=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
success_result = conn.invoke(FunctionName='testFunction', InvocationType='RequestResponse',
|
||||||
|
Payload=json.dumps({'msg': 'So long and thanks for all the fish'}))
|
||||||
|
success_result["StatusCode"].should.equal(202)
|
||||||
|
|
||||||
|
#nasty hack - hope someone has better solution dealing with unicode tests working for Py2 and Py3.
|
||||||
|
base64.b64decode(success_result["LogResult"]).decode('utf-8').replace("u'", "'").should.equal("({'msg': 'So long and thanks for all the fish'}, {})\n\n")
|
||||||
|
|
||||||
|
@mock_ec2
|
||||||
|
@mock_lambda
|
||||||
|
@freeze_time('2015-01-01 00:00:00')
|
||||||
|
def test_invoke_function_get_ec2_volume():
|
||||||
|
conn = boto3.resource("ec2", "us-west-2")
|
||||||
|
vol = conn.create_volume(Size=99, AvailabilityZone='us-west-2')
|
||||||
|
vol = conn.Volume(vol.id)
|
||||||
|
|
||||||
|
conn = boto3.client('lambda', 'us-west-2')
|
||||||
|
conn.create_function(
|
||||||
|
FunctionName='testFunction',
|
||||||
|
Runtime='python2.7',
|
||||||
|
Role='test-iam-role',
|
||||||
|
Handler='lambda_function.handler',
|
||||||
|
Code={
|
||||||
|
'ZipFile': get_test_zip_file2(),
|
||||||
|
},
|
||||||
|
Description='test lambda function',
|
||||||
|
Timeout=3,
|
||||||
|
MemorySize=128,
|
||||||
|
Publish=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
import json
|
||||||
|
success_result = conn.invoke(FunctionName='testFunction', InvocationType='RequestResponse', Payload=json.dumps({'volume_id': vol.id}))
|
||||||
|
success_result["StatusCode"].should.equal(202)
|
||||||
|
|
||||||
import base64
|
import base64
|
||||||
base64.b64decode(success_result["LogResult"]).decode('utf-8').should.equal("Some log file output...")
|
msg = 'get volume details for %s\nVolume - %s state=%s, size=%s\nNone\n\n' % (vol.id, vol.id, vol.state, vol.size)
|
||||||
|
# yet again hacky solution to allow code to run tests for python2 and python3 - pls someone fix :(
|
||||||
|
base64.b64decode(success_result["LogResult"]).decode('utf-8').replace("u'", "'").should.equal(msg)
|
||||||
|
|
||||||
|
|
||||||
@mock_lambda
|
@mock_lambda
|
||||||
@ -100,8 +167,8 @@ def test_create_based_on_s3_with_missing_bucket():
|
|||||||
def test_create_function_from_aws_bucket():
|
def test_create_function_from_aws_bucket():
|
||||||
s3_conn = boto3.client('s3', 'us-west-2')
|
s3_conn = boto3.client('s3', 'us-west-2')
|
||||||
s3_conn.create_bucket(Bucket='test-bucket')
|
s3_conn.create_bucket(Bucket='test-bucket')
|
||||||
|
zip_content = get_test_zip_file2()
|
||||||
|
|
||||||
zip_content = get_test_zip_file()
|
|
||||||
s3_conn.put_object(Bucket='test-bucket', Key='test.zip', Body=zip_content)
|
s3_conn.put_object(Bucket='test-bucket', Key='test.zip', Body=zip_content)
|
||||||
conn = boto3.client('lambda', 'us-west-2')
|
conn = boto3.client('lambda', 'us-west-2')
|
||||||
|
|
||||||
@ -123,7 +190,8 @@ def test_create_function_from_aws_bucket():
|
|||||||
"SubnetIds": ["subnet-123abc"],
|
"SubnetIds": ["subnet-123abc"],
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
result['ResponseMetadata'].pop('HTTPHeaders', None) # this is hard to match against, so remove it
|
result['ResponseMetadata'].pop('HTTPHeaders', None) # this is hard to match against, so remove it
|
||||||
|
result['ResponseMetadata'].pop('RetryAttempts', None) # Botocore inserts retry attempts not seen in Python27
|
||||||
result.should.equal({
|
result.should.equal({
|
||||||
'FunctionName': 'testFunction',
|
'FunctionName': 'testFunction',
|
||||||
'FunctionArn': 'arn:aws:lambda:123456789012:function:testFunction',
|
'FunctionArn': 'arn:aws:lambda:123456789012:function:testFunction',
|
||||||
@ -142,7 +210,6 @@ def test_create_function_from_aws_bucket():
|
|||||||
"SubnetIds": ["subnet-123abc"],
|
"SubnetIds": ["subnet-123abc"],
|
||||||
"VpcId": "vpc-123abc"
|
"VpcId": "vpc-123abc"
|
||||||
},
|
},
|
||||||
|
|
||||||
'ResponseMetadata': {'HTTPStatusCode': 201},
|
'ResponseMetadata': {'HTTPStatusCode': 201},
|
||||||
})
|
})
|
||||||
|
|
||||||
@ -151,8 +218,7 @@ def test_create_function_from_aws_bucket():
|
|||||||
@freeze_time('2015-01-01 00:00:00')
|
@freeze_time('2015-01-01 00:00:00')
|
||||||
def test_create_function_from_zipfile():
|
def test_create_function_from_zipfile():
|
||||||
conn = boto3.client('lambda', 'us-west-2')
|
conn = boto3.client('lambda', 'us-west-2')
|
||||||
|
zip_content = get_test_zip_file1()
|
||||||
zip_content = get_test_zip_file()
|
|
||||||
result = conn.create_function(
|
result = conn.create_function(
|
||||||
FunctionName='testFunction',
|
FunctionName='testFunction',
|
||||||
Runtime='python2.7',
|
Runtime='python2.7',
|
||||||
@ -166,7 +232,9 @@ def test_create_function_from_zipfile():
|
|||||||
MemorySize=128,
|
MemorySize=128,
|
||||||
Publish=True,
|
Publish=True,
|
||||||
)
|
)
|
||||||
result['ResponseMetadata'].pop('HTTPHeaders', None) # this is hard to match against, so remove it
|
result['ResponseMetadata'].pop('HTTPHeaders', None) # this is hard to match against, so remove it
|
||||||
|
result['ResponseMetadata'].pop('RetryAttempts', None) # Botocore inserts retry attempts not seen in Python27
|
||||||
|
|
||||||
result.should.equal({
|
result.should.equal({
|
||||||
'FunctionName': 'testFunction',
|
'FunctionName': 'testFunction',
|
||||||
'FunctionArn': 'arn:aws:lambda:123456789012:function:testFunction',
|
'FunctionArn': 'arn:aws:lambda:123456789012:function:testFunction',
|
||||||
@ -196,7 +264,7 @@ def test_get_function():
|
|||||||
s3_conn = boto3.client('s3', 'us-west-2')
|
s3_conn = boto3.client('s3', 'us-west-2')
|
||||||
s3_conn.create_bucket(Bucket='test-bucket')
|
s3_conn.create_bucket(Bucket='test-bucket')
|
||||||
|
|
||||||
zip_content = get_test_zip_file()
|
zip_content = get_test_zip_file1()
|
||||||
s3_conn.put_object(Bucket='test-bucket', Key='test.zip', Body=zip_content)
|
s3_conn.put_object(Bucket='test-bucket', Key='test.zip', Body=zip_content)
|
||||||
conn = boto3.client('lambda', 'us-west-2')
|
conn = boto3.client('lambda', 'us-west-2')
|
||||||
|
|
||||||
@ -216,7 +284,8 @@ def test_get_function():
|
|||||||
)
|
)
|
||||||
|
|
||||||
result = conn.get_function(FunctionName='testFunction')
|
result = conn.get_function(FunctionName='testFunction')
|
||||||
result['ResponseMetadata'].pop('HTTPHeaders', None) # this is hard to match against, so remove it
|
result['ResponseMetadata'].pop('HTTPHeaders', None) # this is hard to match against, so remove it
|
||||||
|
result['ResponseMetadata'].pop('RetryAttempts', None) # Botocore inserts retry attempts not seen in Python27
|
||||||
|
|
||||||
result.should.equal({
|
result.should.equal({
|
||||||
"Code": {
|
"Code": {
|
||||||
@ -245,14 +314,13 @@ def test_get_function():
|
|||||||
})
|
})
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
@mock_lambda
|
@mock_lambda
|
||||||
@mock_s3
|
@mock_s3
|
||||||
def test_delete_function():
|
def test_delete_function():
|
||||||
s3_conn = boto3.client('s3', 'us-west-2')
|
s3_conn = boto3.client('s3', 'us-west-2')
|
||||||
s3_conn.create_bucket(Bucket='test-bucket')
|
s3_conn.create_bucket(Bucket='test-bucket')
|
||||||
|
|
||||||
zip_content = get_test_zip_file()
|
zip_content = get_test_zip_file2()
|
||||||
s3_conn.put_object(Bucket='test-bucket', Key='test.zip', Body=zip_content)
|
s3_conn.put_object(Bucket='test-bucket', Key='test.zip', Body=zip_content)
|
||||||
conn = boto3.client('lambda', 'us-west-2')
|
conn = boto3.client('lambda', 'us-west-2')
|
||||||
|
|
||||||
@ -272,7 +340,9 @@ def test_delete_function():
|
|||||||
)
|
)
|
||||||
|
|
||||||
success_result = conn.delete_function(FunctionName='testFunction')
|
success_result = conn.delete_function(FunctionName='testFunction')
|
||||||
success_result['ResponseMetadata'].pop('HTTPHeaders', None) # this is hard to match against, so remove it
|
success_result['ResponseMetadata'].pop('HTTPHeaders', None) # this is hard to match against, so remove it
|
||||||
|
success_result['ResponseMetadata'].pop('RetryAttempts', None) # Botocore inserts retry attempts not seen in Python27
|
||||||
|
|
||||||
success_result.should.equal({'ResponseMetadata': {'HTTPStatusCode': 204}})
|
success_result.should.equal({'ResponseMetadata': {'HTTPStatusCode': 204}})
|
||||||
|
|
||||||
conn.delete_function.when.called_with(FunctionName='testFunctionThatDoesntExist').should.throw(botocore.client.ClientError)
|
conn.delete_function.when.called_with(FunctionName='testFunctionThatDoesntExist').should.throw(botocore.client.ClientError)
|
||||||
@ -289,7 +359,7 @@ def test_list_create_list_get_delete_list():
|
|||||||
s3_conn = boto3.client('s3', 'us-west-2')
|
s3_conn = boto3.client('s3', 'us-west-2')
|
||||||
s3_conn.create_bucket(Bucket='test-bucket')
|
s3_conn.create_bucket(Bucket='test-bucket')
|
||||||
|
|
||||||
zip_content = get_test_zip_file()
|
zip_content = get_test_zip_file2()
|
||||||
s3_conn.put_object(Bucket='test-bucket', Key='test.zip', Body=zip_content)
|
s3_conn.put_object(Bucket='test-bucket', Key='test.zip', Body=zip_content)
|
||||||
conn = boto3.client('lambda', 'us-west-2')
|
conn = boto3.client('lambda', 'us-west-2')
|
||||||
|
|
||||||
@ -337,7 +407,9 @@ def test_list_create_list_get_delete_list():
|
|||||||
conn.list_functions()['Functions'].should.equal([expected_function_result['Configuration']])
|
conn.list_functions()['Functions'].should.equal([expected_function_result['Configuration']])
|
||||||
|
|
||||||
func = conn.get_function(FunctionName='testFunction')
|
func = conn.get_function(FunctionName='testFunction')
|
||||||
func['ResponseMetadata'].pop('HTTPHeaders', None) # this is hard to match against, so remove it
|
func['ResponseMetadata'].pop('HTTPHeaders', None) # this is hard to match against, so remove it
|
||||||
|
func['ResponseMetadata'].pop('RetryAttempts', None) # Botocore inserts retry attempts not seen in Python27
|
||||||
|
|
||||||
func.should.equal(expected_function_result)
|
func.should.equal(expected_function_result)
|
||||||
conn.delete_function(FunctionName='testFunction')
|
conn.delete_function(FunctionName='testFunction')
|
||||||
|
|
||||||
|
@ -1,6 +1,7 @@
|
|||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
import json
|
import json
|
||||||
|
|
||||||
|
import base64
|
||||||
import boto
|
import boto
|
||||||
import boto.cloudformation
|
import boto.cloudformation
|
||||||
import boto.datapipeline
|
import boto.datapipeline
|
||||||
@ -1724,10 +1725,29 @@ def test_datapipeline():
|
|||||||
stack_resources.should.have.length_of(1)
|
stack_resources.should.have.length_of(1)
|
||||||
stack_resources[0].physical_resource_id.should.equal(data_pipelines['pipelineIdList'][0]['id'])
|
stack_resources[0].physical_resource_id.should.equal(data_pipelines['pipelineIdList'][0]['id'])
|
||||||
|
|
||||||
|
def _process_lamda(pfunc):
|
||||||
|
import io
|
||||||
|
import zipfile
|
||||||
|
zip_output = io.BytesIO()
|
||||||
|
zip_file = zipfile.ZipFile(zip_output, 'w', zipfile.ZIP_DEFLATED)
|
||||||
|
zip_file.writestr('lambda_function.zip', pfunc)
|
||||||
|
zip_file.close()
|
||||||
|
zip_output.seek(0)
|
||||||
|
return zip_output.read()
|
||||||
|
|
||||||
|
|
||||||
|
def get_test_zip_file1():
|
||||||
|
pfunc = """
|
||||||
|
def lambda_handler(event, context):
|
||||||
|
return (event, context)
|
||||||
|
"""
|
||||||
|
return _process_lamda(pfunc)
|
||||||
|
|
||||||
|
|
||||||
@mock_cloudformation
|
@mock_cloudformation
|
||||||
@mock_lambda
|
@mock_lambda
|
||||||
def test_lambda_function():
|
def test_lambda_function():
|
||||||
|
# switch this to python as backend lambda only supports python execution.
|
||||||
conn = boto3.client('lambda', 'us-east-1')
|
conn = boto3.client('lambda', 'us-east-1')
|
||||||
template = {
|
template = {
|
||||||
"AWSTemplateFormatVersion": "2010-09-09",
|
"AWSTemplateFormatVersion": "2010-09-09",
|
||||||
@ -1736,22 +1756,15 @@ def test_lambda_function():
|
|||||||
"Type": "AWS::Lambda::Function",
|
"Type": "AWS::Lambda::Function",
|
||||||
"Properties": {
|
"Properties": {
|
||||||
"Code": {
|
"Code": {
|
||||||
"ZipFile": {"Fn::Join": [
|
"ZipFile": base64.b64encode(get_test_zip_file1()).decode('utf-8')
|
||||||
"\n",
|
|
||||||
"""
|
|
||||||
exports.handler = function(event, context) {
|
|
||||||
context.succeed();
|
|
||||||
}
|
|
||||||
""".splitlines()
|
|
||||||
]}
|
|
||||||
},
|
},
|
||||||
"Handler": "index.handler",
|
"Handler": "lambda_function.handler",
|
||||||
"Description": "Test function",
|
"Description": "Test function",
|
||||||
"MemorySize": 128,
|
"MemorySize": 128,
|
||||||
"Role": "test-role",
|
"Role": "test-role",
|
||||||
"Runtime": "nodejs",
|
"Runtime": "python2.7"
|
||||||
}
|
}
|
||||||
},
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1765,10 +1778,10 @@ def test_lambda_function():
|
|||||||
result = conn.list_functions()
|
result = conn.list_functions()
|
||||||
result['Functions'].should.have.length_of(1)
|
result['Functions'].should.have.length_of(1)
|
||||||
result['Functions'][0]['Description'].should.equal('Test function')
|
result['Functions'][0]['Description'].should.equal('Test function')
|
||||||
result['Functions'][0]['Handler'].should.equal('index.handler')
|
result['Functions'][0]['Handler'].should.equal('lambda_function.handler')
|
||||||
result['Functions'][0]['MemorySize'].should.equal(128)
|
result['Functions'][0]['MemorySize'].should.equal(128)
|
||||||
result['Functions'][0]['Role'].should.equal('test-role')
|
result['Functions'][0]['Role'].should.equal('test-role')
|
||||||
result['Functions'][0]['Runtime'].should.equal('nodejs')
|
result['Functions'][0]['Runtime'].should.equal('python2.7')
|
||||||
|
|
||||||
|
|
||||||
@mock_cloudformation
|
@mock_cloudformation
|
||||||
|
@ -8,6 +8,7 @@ from freezegun import freeze_time
|
|||||||
from boto.exception import JSONResponseError
|
from boto.exception import JSONResponseError
|
||||||
from moto import mock_dynamodb2
|
from moto import mock_dynamodb2
|
||||||
from tests.helpers import requires_boto_gte
|
from tests.helpers import requires_boto_gte
|
||||||
|
import botocore
|
||||||
try:
|
try:
|
||||||
from boto.dynamodb2.fields import HashKey
|
from boto.dynamodb2.fields import HashKey
|
||||||
from boto.dynamodb2.table import Table
|
from boto.dynamodb2.table import Table
|
||||||
@ -469,6 +470,7 @@ def test_update_item_set():
|
|||||||
})
|
})
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
@mock_dynamodb2
|
@mock_dynamodb2
|
||||||
def test_failed_overwrite():
|
def test_failed_overwrite():
|
||||||
table = Table.create('messages', schema=[
|
table = Table.create('messages', schema=[
|
||||||
@ -585,6 +587,37 @@ def test_boto3_conditions():
|
|||||||
response['Items'][0].should.equal({"username": "johndoe"})
|
response['Items'][0].should.equal({"username": "johndoe"})
|
||||||
|
|
||||||
|
|
||||||
|
@mock_dynamodb2
|
||||||
|
def test_boto3_put_item_conditions_fails():
|
||||||
|
table = _create_user_table()
|
||||||
|
table.put_item(Item={'username': 'johndoe', 'foo': 'bar'})
|
||||||
|
table.put_item.when.called_with(
|
||||||
|
Item={'username': 'johndoe', 'foo': 'baz'},
|
||||||
|
Expected={
|
||||||
|
'foo': {
|
||||||
|
'ComparisonOperator': 'NE',
|
||||||
|
'AttributeValueList': ['bar']
|
||||||
|
}
|
||||||
|
}).should.throw(botocore.client.ClientError)
|
||||||
|
|
||||||
|
|
||||||
|
@mock_dynamodb2
|
||||||
|
def test_boto3_put_item_conditions_pass():
|
||||||
|
table = _create_user_table()
|
||||||
|
table.put_item(Item={'username': 'johndoe', 'foo': 'bar'})
|
||||||
|
table.put_item(
|
||||||
|
Item={'username': 'johndoe', 'foo': 'baz'},
|
||||||
|
Expected={
|
||||||
|
'foo': {
|
||||||
|
'ComparisonOperator': 'EQ',
|
||||||
|
'AttributeValueList': ['bar']
|
||||||
|
}
|
||||||
|
})
|
||||||
|
returned_item = table.get_item(Key={'username': 'johndoe'})
|
||||||
|
assert dict(returned_item)['Item']['foo'].should.equal("baz")
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
@mock_dynamodb2
|
@mock_dynamodb2
|
||||||
def test_scan_pagination():
|
def test_scan_pagination():
|
||||||
table = _create_user_table()
|
table = _create_user_table()
|
||||||
|
@ -73,6 +73,13 @@ def test_instance_launch_and_terminate():
|
|||||||
instance = reservations[0].instances[0]
|
instance = reservations[0].instances[0]
|
||||||
instance.state.should.equal('terminated')
|
instance.state.should.equal('terminated')
|
||||||
|
|
||||||
|
|
||||||
|
@mock_ec2
|
||||||
|
def test_terminate_empty_instances():
|
||||||
|
conn = boto.connect_ec2('the_key', 'the_secret')
|
||||||
|
conn.terminate_instances.when.called_with([]).should.throw(EC2ResponseError)
|
||||||
|
|
||||||
|
|
||||||
@freeze_time("2014-01-01 05:00:00")
|
@freeze_time("2014-01-01 05:00:00")
|
||||||
@mock_ec2
|
@mock_ec2
|
||||||
def test_instance_attach_volume():
|
def test_instance_attach_volume():
|
||||||
@ -330,6 +337,7 @@ def test_get_instances_filtering_by_tag():
|
|||||||
reservations[0].instances[0].id.should.equal(instance1.id)
|
reservations[0].instances[0].id.should.equal(instance1.id)
|
||||||
reservations[0].instances[1].id.should.equal(instance3.id)
|
reservations[0].instances[1].id.should.equal(instance3.id)
|
||||||
|
|
||||||
|
|
||||||
@mock_ec2
|
@mock_ec2
|
||||||
def test_get_instances_filtering_by_tag_value():
|
def test_get_instances_filtering_by_tag_value():
|
||||||
conn = boto.connect_ec2()
|
conn = boto.connect_ec2()
|
||||||
|
@ -87,6 +87,60 @@ def test_create_stream():
|
|||||||
})
|
})
|
||||||
|
|
||||||
|
|
||||||
|
@mock_kinesis
|
||||||
|
@freeze_time("2015-03-01")
|
||||||
|
def test_create_stream_without_redshift():
|
||||||
|
client = boto3.client('firehose', region_name='us-east-1')
|
||||||
|
|
||||||
|
response = client.create_delivery_stream(
|
||||||
|
DeliveryStreamName="stream1",
|
||||||
|
S3DestinationConfiguration={
|
||||||
|
'RoleARN': 'arn:aws:iam::123456789012:role/firehose_delivery_role',
|
||||||
|
'BucketARN': 'arn:aws:s3:::kinesis-test',
|
||||||
|
'Prefix': 'myFolder/',
|
||||||
|
'BufferingHints': {
|
||||||
|
'SizeInMBs': 123,
|
||||||
|
'IntervalInSeconds': 124
|
||||||
|
},
|
||||||
|
'CompressionFormat': 'UNCOMPRESSED',
|
||||||
|
}
|
||||||
|
)
|
||||||
|
stream_arn = response['DeliveryStreamARN']
|
||||||
|
|
||||||
|
response = client.describe_delivery_stream(DeliveryStreamName='stream1')
|
||||||
|
stream_description = response['DeliveryStreamDescription']
|
||||||
|
|
||||||
|
# Sure and Freezegun don't play nicely together
|
||||||
|
created = stream_description.pop('CreateTimestamp')
|
||||||
|
last_updated = stream_description.pop('LastUpdateTimestamp')
|
||||||
|
from dateutil.tz import tzlocal
|
||||||
|
assert created == datetime.datetime(2015, 3, 1, tzinfo=tzlocal())
|
||||||
|
assert last_updated == datetime.datetime(2015, 3, 1, tzinfo=tzlocal())
|
||||||
|
|
||||||
|
stream_description.should.equal({
|
||||||
|
'DeliveryStreamName': 'stream1',
|
||||||
|
'DeliveryStreamARN': stream_arn,
|
||||||
|
'DeliveryStreamStatus': 'ACTIVE',
|
||||||
|
'VersionId': 'string',
|
||||||
|
'Destinations': [
|
||||||
|
{
|
||||||
|
'DestinationId': 'string',
|
||||||
|
'S3DestinationDescription': {
|
||||||
|
'RoleARN': 'arn:aws:iam::123456789012:role/firehose_delivery_role',
|
||||||
|
'RoleARN': 'arn:aws:iam::123456789012:role/firehose_delivery_role',
|
||||||
|
'BucketARN': 'arn:aws:s3:::kinesis-test',
|
||||||
|
'Prefix': 'myFolder/',
|
||||||
|
'BufferingHints': {
|
||||||
|
'SizeInMBs': 123,
|
||||||
|
'IntervalInSeconds': 124
|
||||||
|
},
|
||||||
|
'CompressionFormat': 'UNCOMPRESSED',
|
||||||
|
}
|
||||||
|
},
|
||||||
|
],
|
||||||
|
"HasMoreDestinations": False,
|
||||||
|
})
|
||||||
|
|
||||||
@mock_kinesis
|
@mock_kinesis
|
||||||
@freeze_time("2015-03-01")
|
@freeze_time("2015-03-01")
|
||||||
def test_deescribe_non_existant_stream():
|
def test_deescribe_non_existant_stream():
|
||||||
|
@ -30,6 +30,39 @@ def test_describe_key():
|
|||||||
key['KeyMetadata']['KeyUsage'].should.equal("ENCRYPT_DECRYPT")
|
key['KeyMetadata']['KeyUsage'].should.equal("ENCRYPT_DECRYPT")
|
||||||
|
|
||||||
|
|
||||||
|
@mock_kms
|
||||||
|
def test_describe_key_via_alias():
|
||||||
|
conn = boto.kms.connect_to_region("us-west-2")
|
||||||
|
key = conn.create_key(policy="my policy", description="my key", key_usage='ENCRYPT_DECRYPT')
|
||||||
|
conn.create_alias(alias_name='alias/my-key-alias', target_key_id=key['KeyMetadata']['KeyId'])
|
||||||
|
|
||||||
|
alias_key = conn.describe_key('alias/my-key-alias')
|
||||||
|
alias_key['KeyMetadata']['Description'].should.equal("my key")
|
||||||
|
alias_key['KeyMetadata']['KeyUsage'].should.equal("ENCRYPT_DECRYPT")
|
||||||
|
alias_key['KeyMetadata']['Arn'].should.equal(key['KeyMetadata']['Arn'])
|
||||||
|
|
||||||
|
|
||||||
|
@mock_kms
|
||||||
|
def test_describe_key_via_alias_not_found():
|
||||||
|
conn = boto.kms.connect_to_region("us-west-2")
|
||||||
|
key = conn.create_key(policy="my policy", description="my key", key_usage='ENCRYPT_DECRYPT')
|
||||||
|
conn.create_alias(alias_name='alias/my-key-alias', target_key_id=key['KeyMetadata']['KeyId'])
|
||||||
|
|
||||||
|
conn.describe_key.when.called_with('alias/not-found-alias').should.throw(JSONResponseError)
|
||||||
|
|
||||||
|
|
||||||
|
@mock_kms
|
||||||
|
def test_describe_key_via_arn():
|
||||||
|
conn = boto.kms.connect_to_region("us-west-2")
|
||||||
|
key = conn.create_key(policy="my policy", description="my key", key_usage='ENCRYPT_DECRYPT')
|
||||||
|
arn = key['KeyMetadata']['Arn']
|
||||||
|
|
||||||
|
the_key = conn.describe_key(arn)
|
||||||
|
the_key['KeyMetadata']['Description'].should.equal("my key")
|
||||||
|
the_key['KeyMetadata']['KeyUsage'].should.equal("ENCRYPT_DECRYPT")
|
||||||
|
the_key['KeyMetadata']['KeyId'].should.equal(key['KeyMetadata']['KeyId'])
|
||||||
|
|
||||||
|
|
||||||
@mock_kms
|
@mock_kms
|
||||||
def test_describe_missing_key():
|
def test_describe_missing_key():
|
||||||
conn = boto.kms.connect_to_region("us-west-2")
|
conn = boto.kms.connect_to_region("us-west-2")
|
||||||
@ -58,6 +91,18 @@ def test_enable_key_rotation():
|
|||||||
|
|
||||||
conn.get_key_rotation_status(key_id)['KeyRotationEnabled'].should.equal(True)
|
conn.get_key_rotation_status(key_id)['KeyRotationEnabled'].should.equal(True)
|
||||||
|
|
||||||
|
@mock_kms
|
||||||
|
def test_enable_key_rotation_via_arn():
|
||||||
|
conn = boto.kms.connect_to_region("us-west-2")
|
||||||
|
|
||||||
|
key = conn.create_key(policy="my policy", description="my key", key_usage='ENCRYPT_DECRYPT')
|
||||||
|
key_id = key['KeyMetadata']['Arn']
|
||||||
|
|
||||||
|
conn.enable_key_rotation(key_id)
|
||||||
|
|
||||||
|
conn.get_key_rotation_status(key_id)['KeyRotationEnabled'].should.equal(True)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
@mock_kms
|
@mock_kms
|
||||||
def test_enable_key_rotation_with_missing_key():
|
def test_enable_key_rotation_with_missing_key():
|
||||||
@ -65,6 +110,18 @@ def test_enable_key_rotation_with_missing_key():
|
|||||||
conn.enable_key_rotation.when.called_with("not-a-key").should.throw(JSONResponseError)
|
conn.enable_key_rotation.when.called_with("not-a-key").should.throw(JSONResponseError)
|
||||||
|
|
||||||
|
|
||||||
|
@mock_kms
|
||||||
|
def test_enable_key_rotation_with_alias_name_should_fail():
|
||||||
|
conn = boto.kms.connect_to_region("us-west-2")
|
||||||
|
key = conn.create_key(policy="my policy", description="my key", key_usage='ENCRYPT_DECRYPT')
|
||||||
|
conn.create_alias(alias_name='alias/my-key-alias', target_key_id=key['KeyMetadata']['KeyId'])
|
||||||
|
|
||||||
|
alias_key = conn.describe_key('alias/my-key-alias')
|
||||||
|
alias_key['KeyMetadata']['Arn'].should.equal(key['KeyMetadata']['Arn'])
|
||||||
|
|
||||||
|
conn.enable_key_rotation.when.called_with('alias/my-alias').should.throw(JSONResponseError)
|
||||||
|
|
||||||
|
|
||||||
@mock_kms
|
@mock_kms
|
||||||
def test_disable_key_rotation():
|
def test_disable_key_rotation():
|
||||||
conn = boto.kms.connect_to_region("us-west-2")
|
conn = boto.kms.connect_to_region("us-west-2")
|
||||||
@ -121,6 +178,14 @@ def test_get_key_policy():
|
|||||||
policy = conn.get_key_policy(key_id, 'default')
|
policy = conn.get_key_policy(key_id, 'default')
|
||||||
policy['Policy'].should.equal('my policy')
|
policy['Policy'].should.equal('my policy')
|
||||||
|
|
||||||
|
@mock_kms
|
||||||
|
def test_get_key_policy_via_arn():
|
||||||
|
conn = boto.kms.connect_to_region('us-west-2')
|
||||||
|
|
||||||
|
key = conn.create_key(policy='my policy', description='my key1', key_usage='ENCRYPT_DECRYPT')
|
||||||
|
policy = conn.get_key_policy(key['KeyMetadata']['Arn'], 'default')
|
||||||
|
|
||||||
|
policy['Policy'].should.equal('my policy')
|
||||||
|
|
||||||
@mock_kms
|
@mock_kms
|
||||||
def test_put_key_policy():
|
def test_put_key_policy():
|
||||||
@ -134,6 +199,42 @@ def test_put_key_policy():
|
|||||||
policy['Policy'].should.equal('new policy')
|
policy['Policy'].should.equal('new policy')
|
||||||
|
|
||||||
|
|
||||||
|
@mock_kms
|
||||||
|
def test_put_key_policy_via_arn():
|
||||||
|
conn = boto.kms.connect_to_region('us-west-2')
|
||||||
|
|
||||||
|
key = conn.create_key(policy='my policy', description='my key1', key_usage='ENCRYPT_DECRYPT')
|
||||||
|
key_id = key['KeyMetadata']['Arn']
|
||||||
|
|
||||||
|
conn.put_key_policy(key_id, 'default', 'new policy')
|
||||||
|
policy = conn.get_key_policy(key_id, 'default')
|
||||||
|
policy['Policy'].should.equal('new policy')
|
||||||
|
|
||||||
|
|
||||||
|
@mock_kms
|
||||||
|
def test_put_key_policy_via_alias_should_not_update():
|
||||||
|
conn = boto.kms.connect_to_region('us-west-2')
|
||||||
|
|
||||||
|
key = conn.create_key(policy='my policy', description='my key1', key_usage='ENCRYPT_DECRYPT')
|
||||||
|
conn.create_alias(alias_name='alias/my-key-alias', target_key_id=key['KeyMetadata']['KeyId'])
|
||||||
|
|
||||||
|
conn.put_key_policy.when.called_with('alias/my-key-alias', 'default', 'new policy').should.throw(JSONResponseError)
|
||||||
|
|
||||||
|
policy = conn.get_key_policy(key['KeyMetadata']['KeyId'], 'default')
|
||||||
|
policy['Policy'].should.equal('my policy')
|
||||||
|
|
||||||
|
|
||||||
|
@mock_kms
|
||||||
|
def test_put_key_policy():
|
||||||
|
conn = boto.kms.connect_to_region('us-west-2')
|
||||||
|
|
||||||
|
key = conn.create_key(policy='my policy', description='my key1', key_usage='ENCRYPT_DECRYPT')
|
||||||
|
conn.put_key_policy(key['KeyMetadata']['Arn'], 'default', 'new policy')
|
||||||
|
|
||||||
|
policy = conn.get_key_policy(key['KeyMetadata']['KeyId'], 'default')
|
||||||
|
policy['Policy'].should.equal('new policy')
|
||||||
|
|
||||||
|
|
||||||
@mock_kms
|
@mock_kms
|
||||||
def test_list_key_policies():
|
def test_list_key_policies():
|
||||||
conn = boto.kms.connect_to_region('us-west-2')
|
conn = boto.kms.connect_to_region('us-west-2')
|
||||||
|
@ -354,6 +354,8 @@ def test_list_or_change_tags_for_resource_request():
|
|||||||
response['ResourceTagSet']['Tags'].should.contain(tag1)
|
response['ResourceTagSet']['Tags'].should.contain(tag1)
|
||||||
response['ResourceTagSet']['Tags'].should.contain(tag2)
|
response['ResourceTagSet']['Tags'].should.contain(tag2)
|
||||||
|
|
||||||
|
len(response['ResourceTagSet']['Tags']).should.equal(2)
|
||||||
|
|
||||||
# Try to remove the tags
|
# Try to remove the tags
|
||||||
conn.change_tags_for_resource(
|
conn.change_tags_for_resource(
|
||||||
ResourceType='healthcheck',
|
ResourceType='healthcheck',
|
||||||
|
Loading…
x
Reference in New Issue
Block a user