Need help with moto?
Click the “chat” button below for chat support from the developer who created it, or find similar developers for support.

About the developer

spulec
5.3K Stars 1.5K Forks Apache License 2.0 6.2K Commits 231 Opened issues

Description

A library that allows you to easily mock out tests based on AWS infrastructure.

Services available

!
?

Need anything else?

Contributors list

Moto - Mock AWS Services

Join the chat at https://gitter.im/awsmoto/Lobby

Build Status Coverage Status Docs PyPI PyPI - Python Version PyPI - Downloads Code style: black

Install

To install moto for a specific service:

console
$ pip install moto[ec2,s3]
This will install Moto, and the dependencies required for that specific service.
If you don't care about the number of dependencies, or if you want to mock many AWS services:
console
$ pip install moto[all]

In a nutshell

Moto is a library that allows your tests to easily mock out AWS Services.

Imagine you have the following python code that you want to test:

import boto3

class MyModel(object): def init(self, name, value): self.name = name self.value = value

def save(self):
    s3 = boto3.client('s3', region_name='us-east-1')
    s3.put_object(Bucket='mybucket', Key=self.name, Body=self.value)

Take a minute to think how you would have tested that in the past.

Now see how you could test it with Moto:

import boto3
from moto import mock_s3
from mymodule import MyModel


@mock_s3 def test_my_model_save(): conn = boto3.resource('s3', region_name='us-east-1') # We need to create the bucket since this is all in Moto's 'virtual' AWS account conn.create_bucket(Bucket='mybucket')

model_instance = MyModel('steve', 'is awesome')
model_instance.save()

body = conn.Object('mybucket', 'steve').get()['Body'].read().decode("utf-8")

assert body == 'is awesome'

With the decorator wrapping the test, all the calls to s3 are automatically mocked out. The mock keeps the state of the buckets and keys.

It gets even better! Moto isn't just for Python code and it isn't just for S3. Look at the standalone server mode for more information about running Moto with other languages.
Here's the partial list of the AWS services that currently have support:

| Service Name | Decorator | Comment | |---------------------------|-----------------------|------------------------------| | ACM | @mockacm | | | API Gateway | @mockapigateway | | | Application Autoscaling | @mockapplicationautoscaling | | | Athena | @mockathena | | | Autoscaling | @mockautoscaling | | | Cloudformation | @mockcloudformation | | | Cloudwatch | @mockcloudwatch | | | CloudwatchEvents | @mockevents | | | Cognito Identity | @mockcognitoidentity | | | Cognito Identity Provider | @mockcognitoidp | | | Config | @mockconfig | | | Data Pipeline | @mockdatapipeline | | | DynamoDB | @mockdynamodb | API 20111205. Deprecated. | | DynamoDB2 | @mockdynamodb2 | API 20120810 (Latest) | | EC2 | @mockec2 | | | ECR | @mockecr | | | ECS | @mockecs | | | ELB | @mockelb | | | ELBv2 | @mockelbv2 | | | EMR | @mockemr | | | Forecast | @mockforecast | |
| Glacier | @mock
glacier | | | Glue | @mockglue | | | IAM | @mockiam | | | IoT | @mockiot | | | IoT data | @mockiotdata | | | Kinesis | @mockkinesis | | | KMS | @mockkms | | | Lambda | @mocklambda | Invoking Lambdas requires docker | | Logs | @mocklogs | | | Organizations | @mockorganizations | | | Polly | @mockpolly | | | RAM | @mockram | | | RDS | @mockrds | | | RDS2 | @mockrds2 | | | Redshift | @mockredshift | | | Route53 | @mockroute53 | | | S3 | @mocks3 | | | SecretsManager | @mocksecretsmanager | | | SES | @mockses | | | SNS | @mocksns | | | SQS | @mocksqs | | | SSM | @mockssm | | | Step Functions | @mockstepfunctions | | | STS | @mocksts | | | SWF | @mockswf | | | X-Ray | @mock_xray | |

For a full list of endpoint implementation coverage

Another Example

Imagine you have a function that you use to launch new ec2 instances:

import boto3


def add_servers(ami_id, count): client = boto3.client('ec2', region_name='us-west-1') client.run_instances(ImageId=ami_id, MinCount=count, MaxCount=count)

To test it:

from . import add_servers
from moto import mock_ec2

@mock_ec2 def test_add_servers(): add_servers('ami-1234abcd', 2)

client = boto3.client('ec2', region_name='us-west-1')
instances = client.describe_instances()['Reservations'][0]['Instances']
assert len(instances) == 2
instance1 = instances[0]
assert instance1['ImageId'] == 'ami-1234abcd'

Using moto 1.0.X with boto2

moto 1.0.X mock decorators are defined for boto3 and do not work with boto2. Use the @mockAWSSVCdeprecated to work with boto2.

Using moto with boto2 ```python from moto import mockec2deprecated import boto

@mockec2deprecated def testsomethingwithec2(): ec2conn = boto.ec2.connecttoregion('us-east-1') ec2conn.getonlyinstances(instanceids='i-123456')

When using both boto2 and boto3, one can do this to avoid confusion:
```python
from moto import mock_ec2_deprecated as mock_ec2_b2
from moto import mock_ec2

Usage

All of the services can be used as a decorator, context manager, or in a raw form.

Decorator

@mock_s3
def test_my_model_save():
    # Create Bucket so that test can run
    conn = boto3.resource('s3', region_name='us-east-1')
    conn.create_bucket(Bucket='mybucket')
    model_instance = MyModel('steve', 'is awesome')
    model_instance.save()
    body = conn.Object('mybucket', 'steve').get()['Body'].read().decode()

assert body == 'is awesome'

Context Manager

def test_my_model_save():
    with mock_s3():
        conn = boto3.resource('s3', region_name='us-east-1')
        conn.create_bucket(Bucket='mybucket')
        model_instance = MyModel('steve', 'is awesome')
        model_instance.save()
        body = conn.Object('mybucket', 'steve').get()['Body'].read().decode()

    assert body == 'is awesome'

Raw use

def test_my_model_save():
    mock = mock_s3()
    mock.start()

conn = boto3.resource('s3', region_name='us-east-1')
conn.create_bucket(Bucket='mybucket')

model_instance = MyModel('steve', 'is awesome')
model_instance.save()

assert conn.Object('mybucket', 'steve').get()['Body'].read().decode() == 'is awesome'

mock.stop()

IAM-like Access Control

Moto also has the ability to authenticate and authorize actions, just like it's done by IAM in AWS. This functionality can be enabled by either setting the

INITIAL_NO_AUTH_ACTION_COUNT
environment variable or using the
set_initial_no_auth_action_count
decorator. Note that the current implementation is very basic, see this file for more information.

INITIAL_NO_AUTH_ACTION_COUNT

If this environment variable is set, moto will skip performing any authentication as many times as the variable's value, and only starts authenticating requests afterwards. If it is not set, it defaults to infinity, thus moto will never perform any authentication at all.

set_initial_no_auth_action_count

This is a decorator that works similarly to the environment variable, but the settings are only valid in the function's scope. When the function returns, everything is restored.

@set_initial_no_auth_action_count(4)
@mock_ec2
def test_describe_instances_allowed():
    policy_document = {
        "Version": "2012-10-17",
        "Statement": [
            {
                "Effect": "Allow",
                "Action": "ec2:Describe*",
                "Resource": "*"
            }
        ]
    }
    access_key = ...
    # create access key for an IAM user/assumed role that has the policy above.
    # this part should call __exactly__ 4 AWS actions, so that authentication and authorization starts exactly after this

client = boto3.client('ec2', region_name='us-east-1',
                      aws_access_key_id=access_key['AccessKeyId'],
                      aws_secret_access_key=access_key['SecretAccessKey'])

# if the IAM principal whose access key is used, does not have the permission to describe instances, this will fail
instances = client.describe_instances()['Reservations'][0]['Instances']
assert len(instances) == 0

See the related test suite for more examples.

Experimental: AWS Config Querying

For details about the experimental AWS Config support please see the AWS Config readme here.

Very Important -- Recommended Usage

There are some important caveats to be aware of when using moto:

Failure to follow these guidelines could result in your tests mutating your REAL infrastructure!

How do I avoid tests from mutating my real infrastructure?

You need to ensure that the mocks are actually in place. Changes made to recent versions of

botocore
have altered some of the mock behavior. In short, you need to ensure that you always do the following:
  1. Ensure that your tests have dummy environment variables set up:

    export AWS_ACCESS_KEY_ID='testing'
    export AWS_SECRET_ACCESS_KEY='testing'
    export AWS_SECURITY_TOKEN='testing'
    export AWS_SESSION_TOKEN='testing'
    
  2. VERY IMPORTANT: ensure that you have your mocks set up BEFORE your

    boto3
    client is established. This can typically happen if you import a module that has a
    boto3
    client instantiated outside of a function. See the pesky imports section below on how to work around this.

Example on pytest usage?

If you are a user of pytest, you can leverage pytest fixtures to help set up your mocks and other AWS resources that you would need.

Here is an example: ```python @pytest.fixture(scope='function') def awscredentials(): """Mocked AWS Credentials for moto.""" os.environ['AWSACCESSKEYID'] = 'testing' os.environ['AWSSECRETACCESSKEY'] = 'testing' os.environ['AWSSECURITYTOKEN'] = 'testing' os.environ['AWSSESSION_TOKEN'] = 'testing'

@pytest.fixture(scope='function') def s3(awscredentials): with mocks3(): yield boto3.client('s3', region_name='us-east-1')

@pytest.fixture(scope='function') def sts(awscredentials): with mocksts(): yield boto3.client('sts', region_name='us-east-1')

@pytest.fixture(scope='function') def cloudwatch(awscredentials): with mockcloudwatch(): yield boto3.client('cloudwatch', region_name='us-east-1')

... etc. ```

In the code sample above, all of the AWS/mocked fixtures take in a parameter of

aws_credentials
, which sets the proper fake environment variables. The fake environment variables are used so that
botocore
doesn't try to locate real credentials on your system.

Next, once you need to do anything with the mocked AWS environment, do something like: ```python def testcreatebucket(s3): # s3 is a fixture defined above that yields a boto3 s3 client. # Feel free to instantiate another boto3 S3 client -- Keep note of the region though. s3.create_bucket(Bucket="somebucket")

result = s3.list_buckets()
assert len(result['Buckets']) == 1
assert result['Buckets'][0]['Name'] == 'somebucket'
### Example on unittest usage?

If you use unittest to run tests, and you want to use moto inside setUp or setUpClass, you can do it with .start() and .stop() like:

```python import unittest from moto import mock_s3 import boto3

def func_to_test(bucket_name, key, content): s3 = boto3.resource('s3') object = s3.Object(bucket_name, key) object.put(Body=content)

class MyTest(unittest.TestCase): mock_s3 = mock_s3() bucket_name = 'test-bucket' def setUp(self): self.mock_s3.start()

    # you can use boto3.client('s3') if you prefer
    s3 = boto3.resource('s3')
    bucket = s3.Bucket(self.bucket_name)
    bucket.create(
        CreateBucketConfiguration={
            'LocationConstraint': 'af-south-1'
        })

def tearDown(self):
    self.mock_s3.stop()

def test(self):
    content = b"abc"
    key = '/path/to/obj'

    # run the file which uploads to S3
    func_to_test(self.bucket_name, key, content)

    # check the file was uploaded as expected
    s3 = boto3.resource('s3')
    object = s3.Object(self.bucket_name, key)
    actual = object.get()['Body'].read()
    self.assertEqual(actual, content)

If your test

unittest.TestCase
has only one test method, and you don't need to create AWS resources in
setUp
, you can use the context manager (
with mock_s3():
) within that function, or apply the decorator to that method, instead of
.start()
and
.stop()
. That is simpler, however you then cannot share resource setup code (e.g. S3 bucket creation) between tests.

What about those pesky imports?

Recall earlier, it was mentioned that mocks should be established BEFORE the clients are set up. One way to avoid import issues is to make use of local Python imports -- i.e. import the module inside of the unit test you want to run vs. importing at the top of the file.

Example: ```python def testsomething(s3): from some.package.that.does.something.with.s3 import somefunc # <-- Local import for unit test # ^^ Importing here ensures that the mock has been established.

some_func() # The mock has been established from the "s3" pytest fixture, so this function that uses # a package-level S3 client will properly use the mock and not reach out to AWS. ```

Other caveats

For Tox, Travis CI, and other build systems, you might need to also perform a

touch ~/.aws/credentials
command before running the tests. As long as that file is present (empty preferably) and the environment variables above are set, you should be good to go.

Stand-alone Server Mode

Moto also has a stand-alone server mode. This allows you to utilize the backend structure of Moto even if you don't use Python.

It uses flask, which isn't a default dependency. You can install the server 'extra' package with:

pip install "moto[server]"

You can then start it running a service:

$ moto_server ec2
 * Running on http://127.0.0.1:5000/

You can also pass the port:

$ moto_server ec2 -p3000
 * Running on http://127.0.0.1:3000/

If you want to be able to use the server externally you can pass an IP address to bind to as a hostname or allow any of your external interfaces with 0.0.0.0:

$ moto_server ec2 -H 0.0.0.0
 * Running on http://0.0.0.0:5000/

Please be aware this might allow other network users to access your server.

Then go to localhost to see a list of running instances (it will be empty since you haven't added any yet).

If you want to use boto with this (using the simpler decorators above instead is strongly encouraged), the easiest way is to create a boto config file (

~/.boto
) with the following values:
[Boto]
is_secure = False
https_validate_certificates = False
proxy_port = 5000
proxy = 127.0.0.1

If you want to use boto3 with this, you can pass an

endpoint_url
to the resource
boto3.resource(
    service_name='s3',
    region_name='us-west-1',
    endpoint_url='http://localhost:5000',
)

Caveats

The standalone server has some caveats with some services. The following services require that you update your hosts file for your code to work properly:

  1. s3-control

For the above services, this is required because the hostname is in the form of

AWS_ACCOUNT_ID.localhost
. As a result, you need to add that entry to your host file for your tests to function properly.

Releases

Releases are done from Gitlab Actions. Fairly closely following this: https://packaging.python.org/guides/publishing-package-distribution-releases-using-github-actions-ci-cd-workflows/

  • Commits to
    master
    branch do a dev deploy to pypi.
  • Commits to a tag do a real deploy to pypi.

We use cookies. If you continue to browse the site, you agree to the use of cookies. For more information on our use of cookies please see our Privacy Policy.