Tempest Coding Guide¶
Step 1: Read the OpenStack Style Commandments https://docs.openstack.org/hacking/latest/
Step 2: Read on
Tempest Specific Commandments¶
[T102] Cannot import OpenStack python clients in tempest/api & tempest/scenario tests
[T104] Scenario tests require a services decorator
[T105] Tests cannot use setUpClass/tearDownClass
[T107] Check that a service tag isn’t in the module path
[T108] Check no hyphen at the end of rand_name() argument
[T109] Cannot use testtools.skip decorator; instead use decorators.skip_because from tempest.lib
[T110] Check that service client names of GET should be consistent
[T111] Check that service client names of DELETE should be consistent
[T112] Check that tempest.lib should not import local tempest code
[T113] Check that tests use data_utils.rand_uuid() instead of uuid.uuid4()
[T114] Check that tempest.lib does not use tempest config
[T115] Check that admin tests should exist under admin path
[N322] Method’s default argument shouldn’t be mutable
[T116] Unsupported ‘message’ Exception attribute in PY3
[T117] Check negative tests have
[T118] LOG.warn is deprecated. Enforce use of LOG.warning.
It is recommended to use
tox -eautopep8 before submitting a patch.
Assume nothing about existing test data
Tests should be self contained (provide their own data)
Clean up test data at the completion of each test
Use configuration files for values that will vary by environment
Supported OpenStack Components¶
However, Tempest only offers in-tree integration testing coverage for the following components:
Historically, Tempest offered in-tree testing for other components as well, but since the introduction of the External Plugin Interface, Tempest’s in-tree testing scope has been limited to the projects above. Integration tests for projects not included above should go into one of the relevant plugin projects.
According to the
The Zen of Python the
Errors should never pass silently.
Tempest usually runs in special environment (jenkins gate jobs), in every
error or failure situation we should provide as much error related
information as possible, because we usually do not have the chance to
investigate the situation after the issue happened.
In every test case the abnormal situations must be very verbosely explained, by the exception and the log.
In most cases the very first issue is the most important information.
Try to avoid using
try blocks in the test cases, as both the
finally blocks could replace the original exception,
when the additional operations leads to another exception.
Just letting an exception to propagate, is not a bad idea in a test case, at all.
Try to avoid using any exception handling construct which can hide the errors origin.
If you really need to use a
try block, please ensure the original
exception at least logged. When the exception is logged you usually need
raise the same or a different exception anyway.
self.addCleanup is often a good way to avoid having to catch
exceptions and still ensure resources are correctly cleaned up if the
test fails part way through.
self.assert* methods provided by the unit test framework.
This signals the failures early on.
Avoid using the
self.fail alone, its stack trace will signal
self.fail line as the origin of the error.
Avoid constructing complex boolean expressions for assertion.
self.assertFalse without a
will just tell you the single boolean value, and you will not know anything
about the values used in the formula, the
msg argument might be good enough
for providing more information.
Most other assert method can include more information by default.
self.assertIn can include the whole set.
If the test case fails you can see the related logs and the information carried by the exception (exception class, backtrack and exception info). This and the service logs are your only guide to finding the root cause of flaky issues.
Test cases are independent¶
test_method must be callable individually and MUST NOT depends on,
Test cases MAY depend on commonly initialized resources/facilities, like
credentials management, testresources and so on. These facilities, MUST be able
to work even if just one
test_method is selected for execution.
Service tagging is used to specify which services are exercised by a particular
test method. You specify the services with the
decorator. For example:
Valid service tag names are the same as the list of directories in tempest.api that have tests.
For scenario tests having a service tag is required. For the API tests service
tags are only needed if the test method makes an API call (either directly or
indirectly through another service) that differs from the parent directory
name. For example, any test that make an API call to a service other than Nova
tempest.api.compute would require a service tag for those services,
however they do not need to be tagged as
Tempest leverages test attributes which are a simple but effective way of
distinguishing between different “types” of API tests. A test can be “tagged”
with such attributes using the
decorators.attr decorator, for example:
@decorators.attr(type=['negative']) def test_aggregate_create_aggregate_name_length_less_than_1(self): [...]
These test attributes can be used for test selection via regular expressions.
(?!.*\[.*\bslow\b.*\])(^tempest\.scenario) runs all the tests
scenario test module, except for those tagged with the
attribute (via a negative lookahead in the regular expression). These
attributes are used in Tempest’s
tox.ini as well as Tempest’s Zuul job
definitions for specifying particular batches of Tempest test suites to run.
type='negative' attribute is used to signify that a test is a negative
test, which is a test that handles invalid input gracefully. This attribute
should be applied to all negative test scenarios.
This attribute must be applied to each test that belongs to a negative test class, i.e. a test class name ending with “Negative.*” substring.
type='slow' attribute is used to signify that a test takes a long time
to run, relatively speaking. This attribute is usually applied to
scenario tests, which involve a complicated
series of API operations, the total runtime of which can be relatively long.
This long runtime has performance implications on Zuul jobs, which is why
slow attribute is leveraged to run slow tests on a selective basis,
to keep total Zuul job runtime down to a reasonable time frame.
type='smoke' attribute is used to signify that a test is a so-called
smoke test, which is a type of test that tests the most vital OpenStack
functionality, like listing servers or flavors or creating volumes. The
attribute should be sparingly applied to only the tests that sanity-check the
most essential functionality of an OpenStack cloud.
Test fixtures and resources¶
Test level resources should be cleaned-up after the test execution. Clean-up
is best scheduled using
addCleanup which ensures that the resource cleanup
code is always invoked, and in reverse order with respect to the creation
Test class level resources should be defined in the
of the test class, except for any credential obtained from the credentials
provider, which should be set-up in the
Cleanup is best scheduled using
addClassResourceCleanup which ensures that
the cleanup code is always invoked, and in reverse order with respect to the
In both cases - test level and class level cleanups - a wait loop should be scheduled before the actual delete of resources with an asynchronous delete.
The test base class
BaseTestCase defines Tempest framework for class level
tearDownClass are defined here and cannot be
overwritten by subclasses (enforced via hacking rule T105).
Set-up is split in a series of steps (setup stages), which can be overwritten by test classes. Set-up stages are:
Tear-down is also split in a series of steps (teardown stages), which are stacked for execution only if the corresponding setup stage had been reached during the setup phase. Tear-down stages are:
clear_credentials(defined in the base test class)
Skipping tests should be based on configuration only. If that is not possible, it is likely that either a configuration flag is missing, or the test should fail rather than be skipped. Using discovery for skipping tests is generally discouraged.
When running a test that requires a certain “feature” in the target cloud, if that feature is missing we should fail, because either the test configuration is invalid, or the cloud is broken and the expected “feature” is not there even if the cloud was configured with it.
Error handling is an important aspect of API design and usage. Negative tests are a way to ensure that an application can gracefully handle invalid or unexpected input. However, as a black box integration test suite, Tempest is not suitable for handling all negative test cases, as the wide variety and complexity of negative tests can lead to long test runs and knowledge of internal implementation details. The bulk of negative testing should be handled with project function tests. All negative tests should be based on API-WG guideline . Such negative tests can block any changes from accurate failure code to invalid one.
If facing some gray area which is not clarified on the above guideline, propose a new guideline to the API-WG. With a proposal to the API-WG we will be able to build a consensus across all OpenStack projects and improve the quality and consistency of all the APIs.
In addition, we have some guidelines for additional negative tests.
About BadRequest(HTTP400) case: We can add a single negative tests of BadRequest for each resource and method(POST, PUT). Please don’t implement more negative tests on the same combination of resource and method even if API request parameters are different from the existing test.
About NotFound(HTTP404) case: We can add a single negative tests of NotFound for each resource and method(GET, PUT, DELETE, HEAD). Please don’t implement more negative tests on the same combination of resource and method.
The above guidelines don’t cover all cases and we will grow these guidelines organically over time. Patches outside of the above guidelines are left up to the reviewers’ discretion and if we face some conflicts between reviewers, we will expand the guideline based on our discussion and experience.
Test skips because of Known Bugs¶
If a test is broken because of a bug it is appropriate to skip the test until
bug has been fixed. You should use the
skip_because decorator so that
Tempest’s skip tracking tool can watch the bug status.
@skip_because(bug="980688") def test_this_and_that(self): ...
Do not submit changesets with only testcases which are skipped as they will not be merged.
Consistently check the status code of responses in testcases. The earlier a problem is detected the easier it is to debug, especially where there is complicated setup required.
Parallel Test Execution¶
Tempest by default runs its tests in parallel this creates the possibility for interesting interactions between tests which can cause unexpected failures. Dynamic credentials provides protection from most of the potential race conditions between tests outside the same class. But there are still a few of things to watch out for to try to avoid issues when running your tests in parallel.
Resources outside of a project scope still have the potential to conflict. This is a larger concern for the admin tests since most resources and actions that require admin privileges are outside of projects.
Races between methods in the same class are not a problem because parallelization in Tempest is at the test class level, but if there is a json and xml version of the same test class there could still be a race between methods.
The rand_name() function from tempest.lib.common.utils.data_utils should be used anywhere a resource is created with a name. Static naming should be avoided to prevent resource conflicts.
If the execution of a set of tests is required to be serialized then locking can be used to perform this. See usage of
LockFixturefor examples of using locking. However, LockFixture only helps if you want to separate the execution of two small sets of test cases. On the other hand, if you need to run a set of tests separately from potentially all other tests then
LockFixturedoes not scale as you would need to take the lock in all the other tests too. In this case, you can use the
@serialdecorator on top of the test class holding the tests that need to run separately from the potentially parallel test set. See more in Tempest Test Writing Guide.
Sample Configuration File¶
The sample config file is autogenerated using a script. If any changes are made to the config variables in tempest/config.py then the sample config file must be regenerated. This can be done running:
tox -e genconfig
Unit tests are a separate class of tests in Tempest. They verify Tempest itself, and thus have a different set of guidelines around them:
They can not require anything running externally. All you should need to run the unit tests is the git tree, python and the dependencies installed. This includes running services, a config file, etc.
The unit tests cannot use setUpClass, instead fixtures and testresources should be used for shared state between tests.
For tests being added we need to require inline documentation in the form of docstrings to explain what is being tested. In API tests for a new API a class level docstring should be added to an API reference doc. If one doesn’t exist a TODO comment should be put indicating that the reference needs to be added. For individual API test cases a method level docstring should be used to explain the functionality being tested if the test name isn’t descriptive enough. For example:
def test_get_role_by_id(self): """Get a role by its id."""
the docstring there is superfluous and shouldn’t be added. but for a method like:
def test_volume_backup_create_get_detailed_list_restore_delete(self): pass
a docstring would be useful because while the test title is fairly descriptive the operations being performed are complex enough that a bit more explanation will help people figure out the intent of the test.
For scenario tests a class level docstring describing the steps in the scenario is required. If there is more than one test case in the class individual docstrings for the workflow in each test methods can be used instead. A good example of this would be:
class TestServerBasicOps(manager.ScenarioTest): """The test suite for server basic operations This smoke test case follows this basic set of operations: * Create a keypair for use in launching an instance * Create a security group to control network access in instance * Add simple permissive rules to the security group * Launch an instance * Perform ssh to instance * Verify metadata service * Verify metadata on config_drive * Terminate the instance """
Test Identification with Idempotent ID¶
Every function that provides a test must have an
that is a unique
uuid-4 instance. This ID is used to complement the fully
qualified test name and track test functionality through refactoring. The
format of the metadata looks like:
@decorators.idempotent_id('585e934c-448e-43c4-acbf-d06a9b899997') def test_list_servers_with_detail(self): # The created server should be in the detailed list of all servers ...
Tempest.lib includes a
check-uuid tool that will test for the existence
and uniqueness of idempotent_id metadata for every test. If you have
Tempest installed you run the tool against Tempest by calling from the
It can be invoked against any test suite by passing a package name:
check-uuid --package <package_name>
Tests without an
idempotent_id can be automatically fixed by running
the command with the
--fix flag, which will modify the source package
by inserting randomly generated uuids for every test that does not have
check-uuid tool is used as part of the Tempest gate job
to ensure that all tests have an
Branchless Tempest Considerations¶
Starting with the OpenStack Icehouse release Tempest no longer has any stable branches. This is to better ensure API consistency between releases because the API behavior should not change between releases. This means that the stable branches are also gated by the Tempest master branch, which also means that proposed commits to Tempest must work against both the master and all the currently supported stable branches of the projects. As such there are a few special considerations that have to be accounted for when pushing new changes to Tempest.
1. New Tests for new features¶
When adding tests for new features that were not in previous releases of the
projects the new test has to be properly skipped with a feature flag. This can
be just as simple as using the
testtools.skipUnless decorators to check if the required extension (or
discoverable optional API) or feature is enabled or can be as difficult as
adding a new config option to the appropriate section. If there isn’t a method
of selecting the new feature from the config file then there won’t be a
mechanism to disable the test with older stable releases and the new test
won’t be able to merge.
Introduction of a new feature flag requires specifying a default value for the corresponding config option that is appropriate in the latest OpenStack release. Because Tempest is branchless, the feature flag’s default value will need to be overridden to a value that is appropriate in earlier releases in which the feature isn’t available. In DevStack, this can be accomplished by modifying Tempest’s lib installation script for previous branches (because DevStack is branched).
2. Bug fix on core project needing Tempest changes¶
When trying to land a bug fix which changes a tested API you’ll have to use the following procedure:
1. Propose change to the project, get a +2 on the change even with failing 2. Propose skip on Tempest which will only be approved after the corresponding change in the project has a +2 on change 3. Land project change in master and all open stable branches (if required) 4. Land changed test in Tempest
Otherwise the bug fix won’t be able to land in the project.
Handily, Zuul’s cross-repository dependencies. can be leveraged to do without step 2 and to have steps 3 and 4 happen “atomically”. To do that, make the patch written in step 1 to depend (refer to Zuul’s documentation above) on the patch written in step 4. The commit message for the Tempest change should have a link to the Gerrit review that justifies that change.
3. New Tests for existing features¶
If a test is being added for a feature that exists in all the current releases of the projects then the only concern is that the API behavior is the same across all the versions of the project being tested. If the behavior is not consistent the test will not be able to merge.
For new tests being added to Tempest the assumption is that the API being tested is considered stable and adheres to the OpenStack API stability guidelines. If an API is still considered experimental or in development then it should not be tested by Tempest until it is considered stable.