633 lines
21 KiB
Python
633 lines
21 KiB
Python
import base64
|
|
from io import BytesIO
|
|
import json
|
|
import sure # noqa # pylint: disable=unused-import
|
|
import time
|
|
from zipfile import ZipFile, ZIP_DEFLATED
|
|
import zlib
|
|
|
|
import boto3
|
|
from botocore.exceptions import ClientError
|
|
from datetime import datetime
|
|
from moto import mock_logs, mock_lambda, mock_iam, mock_firehose, mock_s3
|
|
from moto import mock_kinesis
|
|
from moto.core.utils import unix_time_millis
|
|
import pytest
|
|
|
|
from tests.markers import requires_docker
|
|
|
|
|
|
@mock_lambda
|
|
@mock_logs
|
|
def test_put_subscription_filter_update():
|
|
# given
|
|
region_name = "us-east-1"
|
|
client_lambda = boto3.client("lambda", region_name)
|
|
client_logs = boto3.client("logs", region_name)
|
|
log_group_name = "/test"
|
|
log_stream_name = "stream"
|
|
client_logs.create_log_group(logGroupName=log_group_name)
|
|
client_logs.create_log_stream(
|
|
logGroupName=log_group_name, logStreamName=log_stream_name
|
|
)
|
|
function_arn = client_lambda.create_function(
|
|
FunctionName="test",
|
|
Runtime="python3.8",
|
|
Role=_get_role_name(region_name),
|
|
Handler="lambda_function.lambda_handler",
|
|
Code={"ZipFile": _get_test_zip_file()},
|
|
Description="test lambda function",
|
|
Timeout=3,
|
|
MemorySize=128,
|
|
Publish=True,
|
|
)["FunctionArn"]
|
|
|
|
# when
|
|
client_logs.put_subscription_filter(
|
|
logGroupName=log_group_name,
|
|
filterName="test",
|
|
filterPattern="",
|
|
destinationArn=function_arn,
|
|
)
|
|
|
|
# then
|
|
response = client_logs.describe_subscription_filters(logGroupName=log_group_name)
|
|
response["subscriptionFilters"].should.have.length_of(1)
|
|
sub_filter = response["subscriptionFilters"][0]
|
|
creation_time = sub_filter["creationTime"]
|
|
creation_time.should.be.a(int)
|
|
sub_filter["destinationArn"] = "arn:aws:lambda:us-east-1:123456789012:function:test"
|
|
sub_filter["distribution"] = "ByLogStream"
|
|
sub_filter["logGroupName"] = "/test"
|
|
sub_filter["filterName"] = "test"
|
|
sub_filter["filterPattern"] = ""
|
|
|
|
# when
|
|
# to update an existing subscription filter the 'filerName' must be identical
|
|
client_logs.put_subscription_filter(
|
|
logGroupName=log_group_name,
|
|
filterName="test",
|
|
filterPattern="[]",
|
|
destinationArn=function_arn,
|
|
)
|
|
|
|
# then
|
|
response = client_logs.describe_subscription_filters(logGroupName=log_group_name)
|
|
response["subscriptionFilters"].should.have.length_of(1)
|
|
sub_filter = response["subscriptionFilters"][0]
|
|
sub_filter["creationTime"].should.equal(creation_time)
|
|
sub_filter["destinationArn"] = "arn:aws:lambda:us-east-1:123456789012:function:test"
|
|
sub_filter["distribution"] = "ByLogStream"
|
|
sub_filter["logGroupName"] = "/test"
|
|
sub_filter["filterName"] = "test"
|
|
sub_filter["filterPattern"] = "[]"
|
|
|
|
# when
|
|
# only one subscription filter can be associated with a log group
|
|
with pytest.raises(ClientError) as e:
|
|
client_logs.put_subscription_filter(
|
|
logGroupName=log_group_name,
|
|
filterName="test-2",
|
|
filterPattern="",
|
|
destinationArn=function_arn,
|
|
)
|
|
|
|
# then
|
|
ex = e.value
|
|
ex.operation_name.should.equal("PutSubscriptionFilter")
|
|
ex.response["ResponseMetadata"]["HTTPStatusCode"].should.equal(400)
|
|
ex.response["Error"]["Code"].should.contain("LimitExceededException")
|
|
ex.response["Error"]["Message"].should.equal("Resource limit exceeded.")
|
|
|
|
|
|
@mock_lambda
|
|
@mock_logs
|
|
@pytest.mark.network
|
|
@requires_docker
|
|
def test_put_subscription_filter_with_lambda():
|
|
# given
|
|
region_name = "us-east-1"
|
|
client_lambda = boto3.client("lambda", region_name)
|
|
client_logs = boto3.client("logs", region_name)
|
|
log_group_name = "/test"
|
|
log_stream_name = "stream"
|
|
client_logs.create_log_group(logGroupName=log_group_name)
|
|
client_logs.create_log_stream(
|
|
logGroupName=log_group_name, logStreamName=log_stream_name
|
|
)
|
|
function_arn = client_lambda.create_function(
|
|
FunctionName="test",
|
|
Runtime="python3.8",
|
|
Role=_get_role_name(region_name),
|
|
Handler="lambda_function.lambda_handler",
|
|
Code={"ZipFile": _get_test_zip_file()},
|
|
Description="test lambda function",
|
|
Timeout=3,
|
|
MemorySize=128,
|
|
Publish=True,
|
|
)["FunctionArn"]
|
|
|
|
# when
|
|
client_logs.put_subscription_filter(
|
|
logGroupName=log_group_name,
|
|
filterName="test",
|
|
filterPattern="",
|
|
destinationArn=function_arn,
|
|
)
|
|
|
|
# then
|
|
response = client_logs.describe_subscription_filters(logGroupName=log_group_name)
|
|
response["subscriptionFilters"].should.have.length_of(1)
|
|
sub_filter = response["subscriptionFilters"][0]
|
|
sub_filter["creationTime"].should.be.a(int)
|
|
sub_filter["destinationArn"] = "arn:aws:lambda:us-east-1:123456789012:function:test"
|
|
sub_filter["distribution"] = "ByLogStream"
|
|
sub_filter["logGroupName"] = "/test"
|
|
sub_filter["filterName"] = "test"
|
|
sub_filter["filterPattern"] = ""
|
|
|
|
# when
|
|
ts_0 = int(unix_time_millis(datetime.utcnow()))
|
|
ts_1 = int(unix_time_millis(datetime.utcnow())) + 10
|
|
client_logs.put_log_events(
|
|
logGroupName=log_group_name,
|
|
logStreamName=log_stream_name,
|
|
logEvents=[
|
|
{"timestamp": ts_0, "message": "test"},
|
|
{"timestamp": ts_1, "message": "test 2"},
|
|
],
|
|
)
|
|
|
|
# then
|
|
msg_showed_up, received_message = _wait_for_log_msg(
|
|
client_logs, "/aws/lambda/test", "awslogs"
|
|
)
|
|
assert (
|
|
msg_showed_up
|
|
), f"CloudWatch log event was not found. All logs: {received_message}"
|
|
|
|
data = json.loads(received_message)["awslogs"]["data"]
|
|
response = json.loads(
|
|
zlib.decompress(base64.b64decode(data), 16 + zlib.MAX_WBITS).decode("utf-8")
|
|
)
|
|
response["messageType"].should.equal("DATA_MESSAGE")
|
|
response["owner"].should.equal("123456789012")
|
|
response["logGroup"].should.equal("/test")
|
|
response["logStream"].should.equal("stream")
|
|
response["subscriptionFilters"].should.equal(["test"])
|
|
log_events = sorted(response["logEvents"], key=lambda log_event: log_event["id"])
|
|
log_events.should.have.length_of(2)
|
|
log_events[0]["id"].should.be.a(int)
|
|
log_events[0]["message"].should.equal("test")
|
|
log_events[0]["timestamp"].should.equal(ts_0)
|
|
log_events[1]["id"].should.be.a(int)
|
|
log_events[1]["message"].should.equal("test 2")
|
|
log_events[1]["timestamp"].should.equal(ts_1)
|
|
|
|
|
|
@mock_lambda
|
|
@mock_logs
|
|
@pytest.mark.network
|
|
@requires_docker
|
|
def test_subscription_filter_applies_to_new_streams():
|
|
# given
|
|
region_name = "us-east-1"
|
|
client_lambda = boto3.client("lambda", region_name)
|
|
client_logs = boto3.client("logs", region_name)
|
|
log_group_name = "/test"
|
|
log_stream_name = "stream"
|
|
client_logs.create_log_group(logGroupName=log_group_name)
|
|
function_arn = client_lambda.create_function(
|
|
FunctionName="test",
|
|
Runtime="python3.8",
|
|
Role=_get_role_name(region_name),
|
|
Handler="lambda_function.lambda_handler",
|
|
Code={"ZipFile": _get_test_zip_file()},
|
|
Description="test lambda function",
|
|
Timeout=3,
|
|
MemorySize=128,
|
|
Publish=True,
|
|
)["FunctionArn"]
|
|
|
|
# when
|
|
client_logs.put_subscription_filter(
|
|
logGroupName=log_group_name,
|
|
filterName="test",
|
|
filterPattern="",
|
|
destinationArn=function_arn,
|
|
)
|
|
client_logs.create_log_stream( # create log stream after subscription filter applied
|
|
logGroupName=log_group_name, logStreamName=log_stream_name
|
|
)
|
|
ts_0 = int(unix_time_millis(datetime.utcnow()))
|
|
ts_1 = int(unix_time_millis(datetime.utcnow())) + 10
|
|
client_logs.put_log_events(
|
|
logGroupName=log_group_name,
|
|
logStreamName=log_stream_name,
|
|
logEvents=[
|
|
{"timestamp": ts_0, "message": "test"},
|
|
{"timestamp": ts_1, "message": "test 2"},
|
|
],
|
|
)
|
|
|
|
# then
|
|
msg_showed_up, received_message = _wait_for_log_msg(
|
|
client_logs, "/aws/lambda/test", "awslogs"
|
|
)
|
|
assert (
|
|
msg_showed_up
|
|
), f"CloudWatch log event was not found. All logs: {received_message}"
|
|
|
|
data = json.loads(received_message)["awslogs"]["data"]
|
|
response = json.loads(
|
|
zlib.decompress(base64.b64decode(data), 16 + zlib.MAX_WBITS).decode("utf-8")
|
|
)
|
|
response["messageType"].should.equal("DATA_MESSAGE")
|
|
response["owner"].should.equal("123456789012")
|
|
response["logGroup"].should.equal("/test")
|
|
response["logStream"].should.equal("stream")
|
|
response["subscriptionFilters"].should.equal(["test"])
|
|
log_events = sorted(response["logEvents"], key=lambda log_event: log_event["id"])
|
|
log_events.should.have.length_of(2)
|
|
log_events[0]["id"].should.be.a(int)
|
|
log_events[0]["message"].should.equal("test")
|
|
log_events[0]["timestamp"].should.equal(ts_0)
|
|
log_events[1]["id"].should.be.a(int)
|
|
log_events[1]["message"].should.equal("test 2")
|
|
log_events[1]["timestamp"].should.equal(ts_1)
|
|
|
|
|
|
@mock_s3
|
|
@mock_firehose
|
|
@mock_logs
|
|
@pytest.mark.network
|
|
def test_put_subscription_filter_with_firehose():
|
|
# given
|
|
region_name = "us-east-1"
|
|
client_firehose = boto3.client("firehose", region_name)
|
|
client_logs = boto3.client("logs", region_name)
|
|
|
|
log_group_name = "/firehose-test"
|
|
log_stream_name = "delivery-stream"
|
|
client_logs.create_log_group(logGroupName=log_group_name)
|
|
client_logs.create_log_stream(
|
|
logGroupName=log_group_name, logStreamName=log_stream_name
|
|
)
|
|
|
|
# Create a S3 bucket.
|
|
bucket_name = "firehosetestbucket"
|
|
s3_client = boto3.client("s3", region_name=region_name)
|
|
s3_client.create_bucket(
|
|
Bucket=bucket_name,
|
|
CreateBucketConfiguration={"LocationConstraint": "us-west-1"},
|
|
)
|
|
|
|
# Create the Firehose delivery stream that uses that S3 bucket as
|
|
# the destination.
|
|
delivery_stream_name = "firehose_log_test"
|
|
firehose_arn = client_firehose.create_delivery_stream(
|
|
DeliveryStreamName=delivery_stream_name,
|
|
ExtendedS3DestinationConfiguration={
|
|
"RoleARN": _get_role_name(region_name),
|
|
"BucketARN": f"arn:aws:s3::{bucket_name}",
|
|
},
|
|
)["DeliveryStreamARN"]
|
|
|
|
# when
|
|
client_logs.put_subscription_filter(
|
|
logGroupName=log_group_name,
|
|
filterName="firehose-test",
|
|
filterPattern="",
|
|
destinationArn=firehose_arn,
|
|
)
|
|
|
|
# then
|
|
response = client_logs.describe_subscription_filters(logGroupName=log_group_name)
|
|
response["subscriptionFilters"].should.have.length_of(1)
|
|
_filter = response["subscriptionFilters"][0]
|
|
_filter["creationTime"].should.be.a(int)
|
|
_filter["destinationArn"] = firehose_arn
|
|
_filter["distribution"] = "ByLogStream"
|
|
_filter["logGroupName"] = "/firehose-test"
|
|
_filter["filterName"] = "firehose-test"
|
|
_filter["filterPattern"] = ""
|
|
|
|
# when
|
|
ts_0 = int(unix_time_millis(datetime.utcnow()))
|
|
ts_1 = int(unix_time_millis(datetime.utcnow()))
|
|
client_logs.put_log_events(
|
|
logGroupName=log_group_name,
|
|
logStreamName=log_stream_name,
|
|
logEvents=[
|
|
{"timestamp": ts_0, "message": "test"},
|
|
{"timestamp": ts_1, "message": "test 2"},
|
|
],
|
|
)
|
|
|
|
# then
|
|
bucket_objects = s3_client.list_objects_v2(Bucket=bucket_name)
|
|
message = s3_client.get_object(
|
|
Bucket=bucket_name, Key=bucket_objects["Contents"][0]["Key"]
|
|
)
|
|
response = json.loads(
|
|
zlib.decompress(message["Body"].read(), 16 + zlib.MAX_WBITS).decode("utf-8")
|
|
)
|
|
|
|
response["messageType"].should.equal("DATA_MESSAGE")
|
|
response["owner"].should.equal("123456789012")
|
|
response["logGroup"].should.equal("/firehose-test")
|
|
response["logStream"].should.equal("delivery-stream")
|
|
response["subscriptionFilters"].should.equal(["firehose-test"])
|
|
log_events = sorted(response["logEvents"], key=lambda log_event: log_event["id"])
|
|
log_events.should.have.length_of(2)
|
|
log_events[0]["id"].should.be.a(int)
|
|
log_events[0]["message"].should.equal("test")
|
|
log_events[0]["timestamp"].should.equal(ts_0)
|
|
log_events[1]["id"].should.be.a(int)
|
|
log_events[1]["message"].should.equal("test 2")
|
|
log_events[1]["timestamp"].should.equal(ts_1)
|
|
|
|
|
|
@mock_iam
|
|
@mock_logs
|
|
@mock_kinesis
|
|
def test_put_subscription_filter_with_kinesis():
|
|
logs = boto3.client("logs", "ap-southeast-2")
|
|
logs.create_log_group(logGroupName="lg1")
|
|
logs.create_log_stream(logGroupName="lg1", logStreamName="ls1")
|
|
|
|
# Create a DataStream
|
|
kinesis = boto3.client("kinesis", "ap-southeast-2")
|
|
kinesis.create_stream(
|
|
StreamName="test-stream",
|
|
ShardCount=1,
|
|
StreamModeDetails={"StreamMode": "ON_DEMAND"},
|
|
)
|
|
kinesis_datastream = kinesis.describe_stream(StreamName="test-stream")[
|
|
"StreamDescription"
|
|
]
|
|
kinesis_datastream_arn = kinesis_datastream["StreamARN"]
|
|
|
|
# Subscribe to new log events
|
|
logs.put_subscription_filter(
|
|
logGroupName="lg1",
|
|
filterName="kinesis_erica_core_components_v2",
|
|
filterPattern='- "cwlogs.push.publisher"',
|
|
destinationArn=kinesis_datastream_arn,
|
|
distribution="ByLogStream",
|
|
)
|
|
|
|
# Create new log events
|
|
ts_0 = int(unix_time_millis(datetime.utcnow()))
|
|
ts_1 = int(unix_time_millis(datetime.utcnow()))
|
|
logs.put_log_events(
|
|
logGroupName="lg1",
|
|
logStreamName="ls1",
|
|
logEvents=[
|
|
{"timestamp": ts_0, "message": "test"},
|
|
{"timestamp": ts_1, "message": "test 2"},
|
|
],
|
|
)
|
|
|
|
# Verify that Kinesis Stream has this data
|
|
nr_of_records = 0
|
|
# We don't know to which shard it was send, so check all of them
|
|
for shard in kinesis_datastream["Shards"]:
|
|
shard_id = shard["ShardId"]
|
|
|
|
shard_iterator = kinesis.get_shard_iterator(
|
|
StreamName="test-stream", ShardId=shard_id, ShardIteratorType="TRIM_HORIZON"
|
|
)["ShardIterator"]
|
|
|
|
resp = kinesis.get_records(ShardIterator=shard_iterator)
|
|
nr_of_records = nr_of_records + len(resp["Records"])
|
|
assert nr_of_records == 1
|
|
|
|
|
|
@mock_lambda
|
|
@mock_logs
|
|
def test_delete_subscription_filter():
|
|
# given
|
|
region_name = "us-east-1"
|
|
client_lambda = boto3.client("lambda", region_name)
|
|
client_logs = boto3.client("logs", region_name)
|
|
log_group_name = "/test"
|
|
client_logs.create_log_group(logGroupName=log_group_name)
|
|
function_arn = client_lambda.create_function(
|
|
FunctionName="test",
|
|
Runtime="python3.8",
|
|
Role=_get_role_name(region_name),
|
|
Handler="lambda_function.lambda_handler",
|
|
Code={"ZipFile": _get_test_zip_file()},
|
|
Description="test lambda function",
|
|
Timeout=3,
|
|
MemorySize=128,
|
|
Publish=True,
|
|
)["FunctionArn"]
|
|
client_logs.put_subscription_filter(
|
|
logGroupName=log_group_name,
|
|
filterName="test",
|
|
filterPattern="",
|
|
destinationArn=function_arn,
|
|
)
|
|
|
|
# when
|
|
client_logs.delete_subscription_filter(logGroupName="/test", filterName="test")
|
|
|
|
# then
|
|
response = client_logs.describe_subscription_filters(logGroupName=log_group_name)
|
|
response["subscriptionFilters"].should.have.length_of(0)
|
|
|
|
|
|
@mock_lambda
|
|
@mock_logs
|
|
def test_delete_subscription_filter_errors():
|
|
# given
|
|
region_name = "us-east-1"
|
|
client_lambda = boto3.client("lambda", region_name)
|
|
client_logs = boto3.client("logs", region_name)
|
|
log_group_name = "/test"
|
|
client_logs.create_log_group(logGroupName=log_group_name)
|
|
function_arn = client_lambda.create_function(
|
|
FunctionName="test",
|
|
Runtime="python3.8",
|
|
Role=_get_role_name(region_name),
|
|
Handler="lambda_function.lambda_handler",
|
|
Code={"ZipFile": _get_test_zip_file()},
|
|
Description="test lambda function",
|
|
Timeout=3,
|
|
MemorySize=128,
|
|
Publish=True,
|
|
)["FunctionArn"]
|
|
client_logs.put_subscription_filter(
|
|
logGroupName=log_group_name,
|
|
filterName="test",
|
|
filterPattern="",
|
|
destinationArn=function_arn,
|
|
)
|
|
|
|
# when
|
|
with pytest.raises(ClientError) as e:
|
|
client_logs.delete_subscription_filter(
|
|
logGroupName="not-existing-log-group", filterName="test"
|
|
)
|
|
|
|
# then
|
|
ex = e.value
|
|
ex.operation_name.should.equal("DeleteSubscriptionFilter")
|
|
ex.response["ResponseMetadata"]["HTTPStatusCode"].should.equal(400)
|
|
ex.response["Error"]["Code"].should.contain("ResourceNotFoundException")
|
|
ex.response["Error"]["Message"].should.equal(
|
|
"The specified log group does not exist"
|
|
)
|
|
|
|
# when
|
|
with pytest.raises(ClientError) as e:
|
|
client_logs.delete_subscription_filter(
|
|
logGroupName="/test", filterName="wrong-filter-name"
|
|
)
|
|
|
|
# then
|
|
ex = e.value
|
|
ex.operation_name.should.equal("DeleteSubscriptionFilter")
|
|
ex.response["ResponseMetadata"]["HTTPStatusCode"].should.equal(400)
|
|
ex.response["Error"]["Code"].should.contain("ResourceNotFoundException")
|
|
ex.response["Error"]["Message"].should.equal(
|
|
"The specified subscription filter does not exist."
|
|
)
|
|
|
|
|
|
@mock_lambda
|
|
@mock_logs
|
|
def test_put_subscription_filter_errors():
|
|
# given
|
|
client_lambda = boto3.client("lambda", "us-east-1")
|
|
function_arn = client_lambda.create_function(
|
|
FunctionName="test",
|
|
Runtime="python3.8",
|
|
Role=_get_role_name("us-east-1"),
|
|
Handler="lambda_function.lambda_handler",
|
|
Code={"ZipFile": _get_test_zip_file()},
|
|
)["FunctionArn"]
|
|
client = boto3.client("logs", "us-east-1")
|
|
log_group_name = "/test"
|
|
client.create_log_group(logGroupName=log_group_name)
|
|
|
|
# when
|
|
with pytest.raises(ClientError) as e:
|
|
client.put_subscription_filter(
|
|
logGroupName="not-existing-log-group",
|
|
filterName="test",
|
|
filterPattern="",
|
|
destinationArn=function_arn,
|
|
)
|
|
|
|
# then
|
|
ex = e.value
|
|
ex.operation_name.should.equal("PutSubscriptionFilter")
|
|
ex.response["ResponseMetadata"]["HTTPStatusCode"].should.equal(400)
|
|
ex.response["Error"]["Code"].should.contain("ResourceNotFoundException")
|
|
ex.response["Error"]["Message"].should.equal(
|
|
"The specified log group does not exist"
|
|
)
|
|
|
|
# when
|
|
with pytest.raises(ClientError) as e:
|
|
client.put_subscription_filter(
|
|
logGroupName="/test",
|
|
filterName="test",
|
|
filterPattern="",
|
|
destinationArn="arn:aws:lambda:us-east-1:123456789012:function:not-existing",
|
|
)
|
|
|
|
# then
|
|
ex = e.value
|
|
ex.operation_name.should.equal("PutSubscriptionFilter")
|
|
ex.response["ResponseMetadata"]["HTTPStatusCode"].should.equal(400)
|
|
ex.response["Error"]["Code"].should.contain("InvalidParameterException")
|
|
ex.response["Error"]["Message"].should.equal(
|
|
"Could not execute the lambda function. "
|
|
"Make sure you have given CloudWatch Logs permission to execute your function."
|
|
)
|
|
|
|
# when
|
|
with pytest.raises(ClientError) as e:
|
|
client.put_subscription_filter(
|
|
logGroupName="/test",
|
|
filterName="test",
|
|
filterPattern="",
|
|
destinationArn="arn:aws:lambda:us-east-1:123456789012:function:not-existing",
|
|
)
|
|
|
|
# then
|
|
ex = e.value
|
|
ex.operation_name.should.equal("PutSubscriptionFilter")
|
|
ex.response["ResponseMetadata"]["HTTPStatusCode"].should.equal(400)
|
|
ex.response["Error"]["Code"].should.contain("InvalidParameterException")
|
|
ex.response["Error"]["Message"].should.equal(
|
|
"Could not execute the lambda function. "
|
|
"Make sure you have given CloudWatch Logs permission to execute your function."
|
|
)
|
|
|
|
# when we pass an unknown kinesis ARN
|
|
with pytest.raises(ClientError) as e:
|
|
client.put_subscription_filter(
|
|
logGroupName="/test",
|
|
filterName="test",
|
|
filterPattern="",
|
|
destinationArn="arn:aws:kinesis:us-east-1:123456789012:stream/unknown-stream",
|
|
)
|
|
|
|
# then
|
|
err = e.value.response["Error"]
|
|
assert err["Code"] == "InvalidParameterException"
|
|
|
|
|
|
def _get_role_name(region_name):
|
|
with mock_iam():
|
|
iam = boto3.client("iam", region_name=region_name)
|
|
try:
|
|
return iam.get_role(RoleName="test-role")["Role"]["Arn"]
|
|
except ClientError:
|
|
return iam.create_role(
|
|
RoleName="test-role", AssumeRolePolicyDocument="test policy", Path="/"
|
|
)["Role"]["Arn"]
|
|
|
|
|
|
def _get_test_zip_file():
|
|
func_str = """
|
|
def lambda_handler(event, context):
|
|
return event
|
|
"""
|
|
|
|
zip_output = BytesIO()
|
|
zip_file = ZipFile(zip_output, "w", ZIP_DEFLATED)
|
|
zip_file.writestr("lambda_function.py", func_str)
|
|
zip_file.close()
|
|
zip_output.seek(0)
|
|
return zip_output.read()
|
|
|
|
|
|
def _wait_for_log_msg(client, log_group_name, expected_msg_part):
|
|
received_messages = []
|
|
start = time.time()
|
|
while (time.time() - start) < 10:
|
|
result = client.describe_log_streams(logGroupName=log_group_name)
|
|
log_streams = result.get("logStreams")
|
|
if not log_streams:
|
|
time.sleep(1)
|
|
continue
|
|
|
|
for log_stream in log_streams:
|
|
result = client.get_log_events(
|
|
logGroupName=log_group_name, logStreamName=log_stream["logStreamName"]
|
|
)
|
|
received_messages.extend(
|
|
[event["message"] for event in result.get("events")]
|
|
)
|
|
for message in received_messages:
|
|
if expected_msg_part in message:
|
|
return True, message
|
|
time.sleep(1)
|
|
return False, received_messages
|