b739bbe0a5
* Fix typo in the documentation (#71701) Fix typo in the documentation: casting instead of casing (cherry picked from commit1a06587f3b
) * Add how to run unit test link in testing_units_modules doc (#71523) * Add how to run unit test link in testing_units_modules * Fix sanity test (cherry picked from commit7a0e545700
) * Fix typo in delveloping_plugins_network (#71737) (cherry picked from commit4bf61f0714
) * Fix broken bullet list (#71728) (cherry picked from commit00ed5b1f2e
) * vmware: Add docs for filters (#71670) Add a scenario guide for filters in VMware documentation Signed-off-by: Abhijeet Kasurde <akasurde@redhat.com> (cherry picked from commit760334303b
) * [docs] add porting guide for DNF GPG validation (#71640) * [docs] add porting guide for DNF GPG validation Change: - This was a breaking change (security fix), but I neglected to add a porting guide entry for it previously. Tickets: - Refs #71537 - Refs #71539 - Refs #71540 - Refs #71541 Signed-off-by: Rick Elrod <rick@elrod.me> * changes from sivel Signed-off-by: Rick Elrod <rick@elrod.me> (cherry picked from commit7a38c470ba
) * Fixed invalid urls inside guide_packet.rst and collections_using.rst (#71705) * Fixed invalid urls inside guide_packet.rst and collections_using.rst * Reverted fix for collections_using.rst (cherry picked from commitc36e939414
) * Update EXAMPLES in package_facts.py documentation (#71838) this module is not limited to rpm , so remove rpm in tasks name (cherry picked from commit7f62b4733d
) * change duplicated label (cherry picked from commitbcfead8e0f
) Co-authored-by: Guillaume Vincent <guillaume@oslab.fr> Co-authored-by: Amin Vakil <info@aminvakil.com> Co-authored-by: Shufeng <fenghhk@gmail.com> Co-authored-by: Evaristo Rojas <evaristo.rojas@islas.org.mx> Co-authored-by: Abhijeet Kasurde <akasurde@redhat.com> Co-authored-by: Rick Elrod <rick@elrod.me> Co-authored-by: Shounak <25407872+shounak1@users.noreply.github.com> Co-authored-by: roumano <roumano@gmail.com>
243 lines
10 KiB
ReStructuredText
243 lines
10 KiB
ReStructuredText
.. _developing_testing:
|
|
|
|
***************
|
|
Testing Ansible
|
|
***************
|
|
|
|
.. contents::
|
|
:local:
|
|
|
|
|
|
Why test your Ansible contributions?
|
|
====================================
|
|
|
|
If you're a developer, one of the most valuable things you can do is to look at GitHub issues and help fix bugs, since bug-fixing is almost always prioritized over feature development. Even for non-developers, helping to test pull requests for bug fixes and features is still immensely valuable.
|
|
|
|
Ansible users who understand how to write playbooks and roles should be able to test their work. GitHub pull requests will automatically run a variety of tests (for example, Shippable) that show bugs in action. However, contributors must also test their work outside of the automated GitHub checks and show evidence of these tests in the PR to ensure that their work will be more likely to be reviewed and merged.
|
|
|
|
Read on to learn how Ansible is tested, how to test your contributions locally, and how to extend testing capabilities.
|
|
|
|
If you want to learn about testing collections, read :ref:`testing_collections`
|
|
|
|
|
|
|
|
Types of tests
|
|
==============
|
|
|
|
At a high level we have the following classifications of tests:
|
|
|
|
:compile:
|
|
* :ref:`testing_compile`
|
|
* Test python code against a variety of Python versions.
|
|
:sanity:
|
|
* :ref:`testing_sanity`
|
|
* Sanity tests are made up of scripts and tools used to perform static code analysis.
|
|
* The primary purpose of these tests is to enforce Ansible coding standards and requirements.
|
|
:integration:
|
|
* :ref:`testing_integration`
|
|
* Functional tests of modules and Ansible core functionality.
|
|
:units:
|
|
* :ref:`testing_units`
|
|
* Tests directly against individual parts of the code base.
|
|
|
|
|
|
If you're a developer, one of the most valuable things you can do is look at the GitHub
|
|
issues list and help fix bugs. We almost always prioritize bug fixing over feature
|
|
development.
|
|
|
|
Even for non developers, helping to test pull requests for bug fixes and features is still
|
|
immensely valuable. Ansible users who understand writing playbooks and roles should be
|
|
able to add integration tests and so GitHub pull requests with integration tests that show
|
|
bugs in action will also be a great way to help.
|
|
|
|
|
|
Testing within GitHub & Shippable
|
|
=================================
|
|
|
|
|
|
Organization
|
|
------------
|
|
|
|
When Pull Requests (PRs) are created they are tested using Shippable, a Continuous Integration (CI) tool. Results are shown at the end of every PR.
|
|
|
|
When Shippable detects an error and it can be linked back to a file that has been modified in the PR then the relevant lines will be added as a GitHub comment. For example::
|
|
|
|
The test `ansible-test sanity --test pep8` failed with the following errors:
|
|
|
|
lib/ansible/modules/network/foo/bar.py:509:17: E265 block comment should start with '# '
|
|
|
|
The test `ansible-test sanity --test validate-modules` failed with the following error:
|
|
lib/ansible/modules/network/foo/bar.py:0:0: E307 version_added should be 2.4. Currently 2.3
|
|
|
|
From the above example we can see that ``--test pep8`` and ``--test validate-modules`` have identified an issue. The commands given allow you to run the same tests locally to ensure you've fixed all issues without having to push your changes to GitHub and wait for Shippable, for example:
|
|
|
|
If you haven't already got Ansible available, use the local checkout by running::
|
|
|
|
source hacking/env-setup
|
|
|
|
Then run the tests detailed in the GitHub comment::
|
|
|
|
ansible-test sanity --test pep8
|
|
ansible-test sanity --test validate-modules
|
|
|
|
If there isn't a GitHub comment stating what's failed you can inspect the results by clicking on the "Details" button under the "checks have failed" message at the end of the PR.
|
|
|
|
Rerunning a failing CI job
|
|
--------------------------
|
|
|
|
Occasionally you may find your PR fails due to a reason unrelated to your change. This could happen for several reasons, including:
|
|
|
|
* a temporary issue accessing an external resource, such as a yum or git repo
|
|
* a timeout creating a virtual machine to run the tests on
|
|
|
|
If either of these issues appear to be the case, you can rerun the Shippable test by:
|
|
|
|
* adding a comment with ``/rebuild`` (full rebuild) or ``/rebuild_failed`` (rebuild only failed CI nodes) to the PR
|
|
* closing and re-opening the PR (full rebuild)
|
|
* making another change to the PR and pushing to GitHub
|
|
|
|
If the issue persists, please contact us in ``#ansible-devel`` on Freenode IRC.
|
|
|
|
|
|
How to test a PR
|
|
================
|
|
|
|
Ideally, code should add tests that prove that the code works. That's not always possible and tests are not always comprehensive, especially when a user doesn't have access to a wide variety of platforms, or is using an API or web service. In these cases, live testing against real equipment can be more valuable than automation that runs against simulated interfaces. In any case, things should always be tested manually the first time as well.
|
|
|
|
Thankfully, helping to test Ansible is pretty straightforward, assuming you are familiar with how Ansible works.
|
|
|
|
Setup: Checking out a Pull Request
|
|
----------------------------------
|
|
|
|
You can do this by:
|
|
|
|
* checking out Ansible
|
|
* fetching the proposed changes into a test branch
|
|
* testing
|
|
* commenting on that particular issue on GitHub
|
|
|
|
Here's how:
|
|
|
|
.. warning::
|
|
Testing source code from GitHub pull requests sent to us does have some inherent risk, as the source code
|
|
sent may have mistakes or malicious code that could have a negative impact on your system. We recommend
|
|
doing all testing on a virtual machine, whether a cloud instance, or locally. Some users like Vagrant
|
|
or Docker for this, but they are optional. It is also useful to have virtual machines of different Linux or
|
|
other flavors, since some features (for example, package managers such as apt or yum) are specific to those OS versions.
|
|
|
|
|
|
Create a fresh area to work::
|
|
|
|
|
|
git clone https://github.com/ansible/ansible.git ansible-pr-testing
|
|
cd ansible-pr-testing
|
|
|
|
Next, find the pull request you'd like to test and make note of its number. It will look something like this::
|
|
|
|
Use os.path.sep instead of hardcoding / #65381
|
|
|
|
.. note:: Only test ``ansible:devel``
|
|
|
|
It is important that the PR request target be ``ansible:devel``, as we do not accept pull requests into any other branch. Dot releases are cherry-picked manually by Ansible staff.
|
|
|
|
Use the pull request number when you fetch the proposed changes and create your branch for testing::
|
|
|
|
git fetch origin refs/pull/XXXX/head:testing_PRXXXX
|
|
git checkout testing_PRXXXX
|
|
|
|
The first command fetches the proposed changes from the pull request and creates a new branch named ``testing_PRXXXX``, where the XXXX is the actual number associated with the pull request (for example, 65381). The second command checks out the newly created branch.
|
|
|
|
.. note::
|
|
If the GitHub user interface shows that the pull request will not merge cleanly, we do not recommend proceeding if you are not somewhat familiar with git and coding, as you will have to resolve a merge conflict. This is the responsibility of the original pull request contributor.
|
|
|
|
.. note::
|
|
Some users do not create feature branches, which can cause problems when they have multiple, unrelated commits in their version of ``devel``. If the source looks like ``someuser:devel``, make sure there is only one commit listed on the pull request.
|
|
|
|
The Ansible source includes a script that allows you to use Ansible directly from source without requiring a
|
|
full installation that is frequently used by developers on Ansible.
|
|
|
|
Simply source it (to use the Linux/Unix terminology) to begin using it immediately::
|
|
|
|
source ./hacking/env-setup
|
|
|
|
This script modifies the ``PYTHONPATH`` environment variables (along with a few other things), which will be temporarily
|
|
set as long as your shell session is open.
|
|
|
|
Testing the Pull Request
|
|
------------------------
|
|
|
|
At this point, you should be ready to begin testing!
|
|
|
|
Some ideas of what to test are:
|
|
|
|
* Create a test Playbook with the examples in and check if they function correctly
|
|
* Test to see if any Python backtraces returned (that's a bug)
|
|
* Test on different operating systems, or against different library versions
|
|
|
|
Run sanity tests
|
|
````````````````
|
|
|
|
.. code:: shell
|
|
|
|
ansible-test sanity
|
|
|
|
More information: :ref:`testing_sanity`
|
|
|
|
Run unit tests
|
|
``````````````
|
|
|
|
.. code:: shell
|
|
|
|
ansible-test units
|
|
|
|
More information: :ref:`testing_units`
|
|
|
|
Run integration tests
|
|
`````````````````````
|
|
|
|
.. code:: shell
|
|
|
|
ansible-test integration -v ping
|
|
|
|
More information: :ref:`testing_integration`
|
|
|
|
Any potential issues should be added as comments on the pull request (and it's acceptable to comment if the feature works as well), remembering to include the output of ``ansible --version``
|
|
|
|
Example::
|
|
|
|
Works for me! Tested on `Ansible 2.3.0`. I verified this on CentOS 6.5 and also Ubuntu 14.04.
|
|
|
|
If the PR does not resolve the issue, or if you see any failures from the unit/integration tests, just include that output instead:
|
|
|
|
| This change causes errors for me.
|
|
|
|
|
| When I ran this Ubuntu 16.04 it failed with the following:
|
|
|
|
|
| \```
|
|
| some output
|
|
| StackTrace
|
|
| some other output
|
|
| \```
|
|
|
|
Code Coverage Online
|
|
````````````````````
|
|
|
|
`The online code coverage reports <https://codecov.io/gh/ansible/ansible>`_ are a good way
|
|
to identify areas for testing improvement in Ansible. By following red colors you can
|
|
drill down through the reports to find files which have no tests at all. Adding both
|
|
integration and unit tests which show clearly how code should work, verify important
|
|
Ansible functions and increase testing coverage in areas where there is none is a valuable
|
|
way to help improve Ansible.
|
|
|
|
The code coverage reports only cover the ``devel`` branch of Ansible where new feature
|
|
development takes place. Pull requests and new code will be missing from the codecov.io
|
|
coverage reports so local reporting is needed. Most ``ansible-test`` commands allow you
|
|
to collect code coverage, this is particularly useful to indicate where to extend
|
|
testing. See :ref:`testing_running_locally` for more information.
|
|
|
|
|
|
Want to know more about testing?
|
|
================================
|
|
|
|
If you'd like to know more about the plans for improving testing Ansible then why not join the
|
|
`Testing Working Group <https://github.com/ansible/community/blob/master/meetings/README.md>`_.
|