* write boilerplate code * generate boilerplate code with scaffold script * create mediapackage channel * remove duplicate mediapackage reference * remove Channel key from mediapackage response * describe_channel endpoint added * create_origin_endpoint-added * keys changed to camel case to fix issue * minor changes to clean up * minor clean up again * implement & test delete_channel * delete origin endpoint created; WIP-tests failing * fix delete_origin_endpoint issue * refactor function call * delete origin endpoint completed; test_server tests added * implement and test describe_origin_endpoint * update origin endpoint added * remove print statements * implement test_list_origin_endpoint_succeeds * create test name changed * create test name changed * changes after flake8 and black run * url assertion added to decribe origin endpoint test * region dependent url enabled * initial commit; WIP * create container added, still WIP * create_container working * test_create_channel_succeeds * test_describe_container_succeeds * get_lifecycle_policy added; error tests added * changes to pass linting * added exception for container but no policy * linting * put_container_policy added * put_metric_policy added * list_containers added * resolved linting * test_describe_container_raises_error_if_container_does_not_exist * added __init__ file * __init__ added to mediapackage as well * Mediastore (#20) * initial commit; WIP * create container added, still WIP * create_container working * test_create_channel_succeeds * test_describe_container_succeeds * get_lifecycle_policy added; error tests added * changes to pass linting * added exception for container but no policy * linting * put_container_policy added * put_metric_policy added * list_containers added * resolved linting * test_describe_container_raises_error_if_container_does_not_exist * added __init__ file * __init__ added to mediapackage as well Co-authored-by: FranckNdame <franck.mpouli@yahoo.com> * test_server fixed; resolved rebasing mix ups on tests * [FIX] Ensures MediaConnect create_flow sets a valid sourceArn * code clean up Co-authored-by: Anya <anya.champaneria@capablue.com> Co-authored-by: AnyaChamp <71766808+AnyaChamp@users.noreply.github.com>
119 lines
4.7 KiB
Python
119 lines
4.7 KiB
Python
from __future__ import unicode_literals
|
|
|
|
import importlib
|
|
|
|
BACKENDS = {
|
|
"acm": ("acm", "acm_backends"),
|
|
"apigateway": ("apigateway", "apigateway_backends"),
|
|
"athena": ("athena", "athena_backends"),
|
|
"applicationautoscaling": (
|
|
"applicationautoscaling",
|
|
"applicationautoscaling_backends",
|
|
),
|
|
"autoscaling": ("autoscaling", "autoscaling_backends"),
|
|
"batch": ("batch", "batch_backends"),
|
|
"cloudformation": ("cloudformation", "cloudformation_backends"),
|
|
"cloudwatch": ("cloudwatch", "cloudwatch_backends"),
|
|
"codecommit": ("codecommit", "codecommit_backends"),
|
|
"codepipeline": ("codepipeline", "codepipeline_backends"),
|
|
"cognito-identity": ("cognitoidentity", "cognitoidentity_backends"),
|
|
"cognito-idp": ("cognitoidp", "cognitoidp_backends"),
|
|
"config": ("config", "config_backends"),
|
|
"datapipeline": ("datapipeline", "datapipeline_backends"),
|
|
"datasync": ("datasync", "datasync_backends"),
|
|
"dynamodb": ("dynamodb", "dynamodb_backends"),
|
|
"dynamodb2": ("dynamodb2", "dynamodb_backends2"),
|
|
"dynamodbstreams": ("dynamodbstreams", "dynamodbstreams_backends"),
|
|
"ec2": ("ec2", "ec2_backends"),
|
|
"ec2instanceconnect": ("ec2instanceconnect", "ec2instanceconnect_backends"),
|
|
"ecr": ("ecr", "ecr_backends"),
|
|
"ecs": ("ecs", "ecs_backends"),
|
|
"elasticbeanstalk": ("elasticbeanstalk", "eb_backends"),
|
|
"elb": ("elb", "elb_backends"),
|
|
"elbv2": ("elbv2", "elbv2_backends"),
|
|
"emr": ("emr", "emr_backends"),
|
|
"events": ("events", "events_backends"),
|
|
"glacier": ("glacier", "glacier_backends"),
|
|
"glue": ("glue", "glue_backends"),
|
|
"iam": ("iam", "iam_backends"),
|
|
"instance_metadata": ("instance_metadata", "instance_metadata_backends"),
|
|
"iot": ("iot", "iot_backends"),
|
|
"iot-data": ("iotdata", "iotdata_backends"),
|
|
"kinesis": ("kinesis", "kinesis_backends"),
|
|
"kms": ("kms", "kms_backends"),
|
|
"lambda": ("awslambda", "lambda_backends"),
|
|
"logs": ("logs", "logs_backends"),
|
|
"managedblockchain": ("managedblockchain", "managedblockchain_backends"),
|
|
"moto_api": ("core", "moto_api_backends"),
|
|
"opsworks": ("opsworks", "opsworks_backends"),
|
|
"organizations": ("organizations", "organizations_backends"),
|
|
"polly": ("polly", "polly_backends"),
|
|
"ram": ("ram", "ram_backends"),
|
|
"rds": ("rds2", "rds2_backends"),
|
|
"redshift": ("redshift", "redshift_backends"),
|
|
"resource-groups": ("resourcegroups", "resourcegroups_backends"),
|
|
"resourcegroupstaggingapi": (
|
|
"resourcegroupstaggingapi",
|
|
"resourcegroupstaggingapi_backends",
|
|
),
|
|
"route53": ("route53", "route53_backends"),
|
|
"s3": ("s3", "s3_backends"),
|
|
"s3bucket_path": ("s3", "s3_backends"),
|
|
"sagemaker": ("sagemaker", "sagemaker_backends"),
|
|
"secretsmanager": ("secretsmanager", "secretsmanager_backends"),
|
|
"ses": ("ses", "ses_backends"),
|
|
"sns": ("sns", "sns_backends"),
|
|
"sqs": ("sqs", "sqs_backends"),
|
|
"ssm": ("ssm", "ssm_backends"),
|
|
"stepfunctions": ("stepfunctions", "stepfunction_backends"),
|
|
"sts": ("sts", "sts_backends"),
|
|
"swf": ("swf", "swf_backends"),
|
|
"transcribe": ("transcribe", "transcribe_backends"),
|
|
"xray": ("xray", "xray_backends"),
|
|
"kinesisvideo": ("kinesisvideo", "kinesisvideo_backends"),
|
|
"medialive": ("medialive", "medialive_backends"),
|
|
"kinesis-video-archived-media": (
|
|
"kinesisvideoarchivedmedia",
|
|
"kinesisvideoarchivedmedia_backends",
|
|
),
|
|
"forecast": ("forecast", "forecast_backends"),
|
|
"support": ("support", "support_backends"),
|
|
"mediaconnect": ("mediaconnect", "mediaconnect_backends"),
|
|
"mediapackage": ("mediapackage", "mediapackage_backends"),
|
|
"mediastore": ("mediastore", "mediastore_backends"),
|
|
}
|
|
|
|
|
|
def _import_backend(module_name, backends_name):
|
|
module = importlib.import_module("moto." + module_name)
|
|
return getattr(module, backends_name)
|
|
|
|
|
|
def backends():
|
|
for module_name, backends_name in BACKENDS.values():
|
|
yield _import_backend(module_name, backends_name)
|
|
|
|
|
|
def named_backends():
|
|
for name, (module_name, backends_name) in BACKENDS.items():
|
|
yield name, _import_backend(module_name, backends_name)
|
|
|
|
|
|
def get_backend(name):
|
|
module_name, backends_name = BACKENDS[name]
|
|
return _import_backend(module_name, backends_name)
|
|
|
|
|
|
def search_backend(predicate):
|
|
for name, backend in named_backends():
|
|
if predicate(backend):
|
|
return name
|
|
|
|
|
|
def get_model(name, region_name):
|
|
for backends_ in backends():
|
|
for region, backend in backends_.items():
|
|
if region == region_name:
|
|
models = getattr(backend.__class__, "__models__", {})
|
|
if name in models:
|
|
return list(getattr(backend, models[name])())
|