Overview of Testing in Tezos ============================ The goal of this document is to give an overview on how testing is done in Tezos, and to help Tezos contributors use the test suite and write tests by pointing them towards the most appropriate testing framework for their use case. Finally, this guide explains how tests can be :ref:`run automatically in the Tezos CI ` and how to :ref:`measure test coverage `. The frameworks used in Tezos can be categorized along two axes: the type of component they test, and the type of testing they perform. We distinguish the following components: - Node - Protocol - Michelson interpreter - Stitching - Networked nodes - Client - Ledger application - Baker Secondly, these components can be tested at different levels of granularity. Additionally, tests can verify functionality, but also non-functional properties such as performance (execution time, memory and disk usage). We distinguish: Unit testing Unit testing tests software units, typically functions, in isolation. Integration testing Integration testing tests compositions of smaller units. System testing System testing tests the final binaries directly. Regression testing In general, regression testing aims to detect the re-introduction of previously identified bugs. It can also refer to a coarse-grained type of testing where the output of a test execution is compared to a pre-recorded log of expected output. We here use "regression testing" to refer to the second meaning. Property testing / Fuzzing Both property testing and fuzzing test code with automatically generated inputs. Property testing is typically used to ensure functional correctness, and gives the user more control over generated input and the expected output. Fuzzing is typically used to search for security weaknesses and often guides input generation with the goal of increasing test coverage. Performance testing Testing of non-functional aspects such as run-time, memory and disk usage. Acceptance testing Testing of the software in real conditions. It is usually slower, more costly and less amenable to automation than integration or system testing. It is often the final step in the testing process and is performed before a release. In Tezos, acceptance testing is done by running a test net. .. Inline testing Inline testing refers to a fine-grained type of testing, where tests are interleaved with the tested code. The inline tests are run when the tested code is executed, and typically removed in production builds. By combining the two axes, we obtain the following matrix. Each cell contains the frameworks appropriate for the corresponding component and testing type. The frameworks are linked to a sub-section of this page where the framework is presented in more detail. .. MT: :ref:`Michelson unit tests `. .. csv-table:: Testing frameworks and their applications in Tezos. PT: :ref:`Python testing and execution framework `, AT: :ref:`alcotest_section`, PBT: :ref:`property_based_test`, FT: :ref:`flextesa_section`, TZ: :ref:`tezt_section` :header: "Component","Unit","Property","Integration","System","Regression" "Node",":ref:`AT `",":ref:`PBT `",":ref:`AT `",":ref:`PT `, :ref:`FT `, :ref:`TZ `" "-- Protocol",":ref:`AT `",":ref:`PBT `","" "-- -- Michelson interpreter",":ref:`AT `","","",":ref:`PT `",":ref:`PT `" "Client","",":ref:`PBT `","",":ref:`PT `, :ref:`FT `, :ref:`TZ `" "Networked nodes","--","",":ref:`PT `, :ref:`FT `","", "" "Endorser","","","",":ref:`FT `" "Baker","","","",":ref:`FT `" Testing frameworks ------------------ .. _alcotest_section: Alcotest ~~~~~~~~ `Alcotest `_ is a library for unit and integration testing in OCaml. Alcotest is the primary tool in Tezos for unit and integration testing of OCaml code. Typical use cases: - Verifying simple input-output specifications for functions with a hard-coded set of input-output pairs. - OCaml integration tests. Example tests: - Unit tests for :src:`src/lib_requester`, in :src:`src/lib_requester/test/test_requester.ml`. To execute them locally, run ``dune build @src/lib_requester/runtest`` in the Tezos root. To execute them on :ref:`your own machine ` using the GitLab CI system, run ``gitlab-runner exec docker unit:requester``. - Integration tests for the P2P layer in the shell. For instance :src:`src/lib_p2p/test/test_p2p_pool.ml`. This test forks a set of processes that exercise large parts of the P2P layer. To execute it locally, run ``dune build @runtest_p2p_pool`` in the Tezos root. To execute the P2P tests on :ref:`your own machine ` using the GitLab CI system, run ``gitlab-runner exec docker unit:p2p``. The job-name ``unit:p2p`` is ill-chosen, since the test is in fact an integration test. References: - `Alcotest README `_. .. _property_based_test: QCheck ~~~~~~~ `QCheck `_ is a library for property-based testing in OCaml. Typical use cases: - Verifying input-output invariants for functions with randomized inputs. Example test: - QCheck is used in :src:`src/lib_base/test/test_time.ml` to test the `Tezos_base.Time `_ module. For instance, subtracting and then adding a random amount of seconds to a random time should give back the original time: this tests that ``add`` and ``diff`` are consistent (and the inverse of each other). To run this test, you need to run ``dune exec src/lib_base/test/test_time.exe``. References: - `QCheck README `_ - `QCheck module documentation `_ .. _pytest_section: Python testing and execution framework ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ The Tezos project uses `pytest `_, a Python testing framework, combined with :doc:`tezos-launchers `, a Python wrapper ``tezos-node`` and ``tezos-client``, to perform integration testing of the node, the client, networks of nodes and daemons such as the baker and endorser. We also use `pytest-regtest `_, a pytest plugin that enables regression testing. Typical use cases: - Testing the commands of ``tezos-client``. This allows to test the full chain: from client, to node RPC to the implementation of the economic protocol. - Test networks of nodes, with daemons. - Detecting unintended changes in the output of a component, using ``pytest-regtest``. Example tests: - Detecting unintended changes in the behavior of the node's Michelson interpreter (in :src:`tests_python/tests_alpha/test_contract_opcodes.py`). To execute it locally, run ``cd tests_python && poetry run pytest tests/test_contract_opcodes.py`` in the Tezos root. To execute them on :ref:`your own machine ` using the GitLab CI system, run ``gitlab-runner exec docker integration:contract_opcodes``. - Setting up networks of nodes and ensuring their connection (in :src:`tests_python/tests_alpha/test_p2p.py`). To execute it locally, run ``cd tests_python && poetry run pytest tests/test_p2p.py`` in the Tezos root. To execute them on :ref:`your own machine ` using the GitLab CI system, run ``gitlab-runner exec docker integration:p2p``. References: - `Pytest Documentation `_ - :doc:`python_testing_framework` - `pytest-regtest README `_ - `pytest-regtest pip package `_ - :ref:`Section in Tezos Developer Documentation on pytest-regtest ` .. _flextesa_section: Flextesa ~~~~~~~~ Flextesa (Flexible Test Sandboxes) is an OCaml library for setting up configurable and scriptable sandboxes to meet specific testing needs. Flextesa can also be used for interactive tests. This is used, for instance, in some tests that require the user to interact with the Ledger application. Typical use cases: - In terms of use cases, Flextesa is similar to the :ref:`Python testing and execution framework `. Example test: - Testing double baking, accusations and double-baking accusation scenarios (in :src:`src/bin_sandbox/command_accusations.ml`) References: - :doc:`Section in Tezos Developer Documentation on Flextesa ` - `Blog post introducing Flextesa `_ - `GitLab repository `_ - `An example setting up a Babylon docker sandbox `_ - `API documentation `_ .. _tezt_section: Tezt ~~~~ :doc:`Tezt ` is a system testing framework for Tezos. It is intended as a replacement to Flextesa and as an OCaml-based alternative to :ref:`Python testing and execution framework `. Like the latter, Tezt is also capable of regression testing. Tezt focuses on tests that run in the CI, although it is also used for some manual tests (see the :src:`tezt/manual_tests` folder). Its main strengths are summarized in its :doc:`section in the Tezos Developer Documentation `. Conceptually Tezt consists of a generic framework for writing tests interacting with external processes, and a set of Tezos-specific modules for interacting with the Tezos binaries: the client, baker, etc. Typical use cases: - In terms of use cases, Tezt is similar to the :ref:`Python testing and execution framework ` and :ref:`Flextesa `. It can be used by authors that prefer OCaml for writing system tests. Example tests: - Testing baking (in :src:`tezt/tests/basic.ml`) - Testing double baking and double endorsement scenarios (in :src:`tezt/tests/double_bake.ml`). This test is a rewrite of the Flextesa double baking scenario mentioned above, that demonstrates the difference between the two frameworks. - Testing absence of regressions in encodings (in :src:`tezt/tests/encoding.ml`) References: - :doc:`Section in Tezos Developer Documentation on Tezt ` - `General API documentation `_ - `Tezos-specific API documentation `_ Long Tests and Performance regression Test Framework """""""""""""""""""""""""""""""""""""""""""""""""""" Tezt is also used for tests that are too long to run in the CI. Those tests are run on dedicated machines and can send data points to an `InfluxDB `__ instance to produce graphs using `Grafana `__ and/or detect performance regressions. See :doc:`long-tezts`. .. .. _michelson_unit_tests: Michelson unit tests -------------------- The `Michelson unit test proposal `__ defines a format for unit tests for Michelson snippets. If the proposal is eventually accepted, then these tests will be executable through ``tezos-client``. Example use cases: - Verifying the functional (input--output) behavior of snippets of Michelson instructions. - Conformance testing for Michelson interpreters. References: - `Merge request defining the Michelson unit test format `_ - `A conformance test suite for Michelson interpreter using the Michelson unit test format `_ .. _gitlab_test_ci: Executing tests locally ----------------------- Whereas executing the tests through the CI, as described below, is the standard and most convenient way of running the full test suite, they can also be executed locally. All tests can be run with ``make test`` in the project root. However, this can take some time, and some tests are resource-intensive or require additional configuration. Alternatively, one can run subsets of tests identified by a specialized target ``test-*``. For instance, ``make test-unit`` runs the alcotest tests and should be quite fast. See the project ``Makefile`` for the full list of testing targets. .. _measuring-test-coverage: Measuring test coverage ~~~~~~~~~~~~~~~~~~~~~~~ We measure `test coverage `_ with `bisect_ppx `_. This tool is used to see which lines in the code source are actually executed when running one or several tests. Importantly, it tells us which parts of the code aren't tested. We describe here how ``bisect_ppx`` can be used locally. See below for usage with CI. To install ``bisect_ppx``, run the following command from the root of the project directory: :: make build-dev-deps The OCaml code should be instrumented in order to generate coverage data. This is done by prepending :: ./scripts/with_coverage.sh to build and test commands run from the root of the project directory. For example, :: ./scripts/with_coverage.sh make ./scripts/with_coverage.sh make test-coverage Generate the HTML report from the coverage files using :: make coverage-report The generated report is available in ``_coverage_report/index.html``. It shows for each file, which lines have been executed at least once, by at least one of the tests. Clean up coverage data (output and report) with: :: make coverage-clean The helper ``./scripts/with_coverage.sh`` can also be used outside make commands (e.g. with ``dune``, ``poetry``). For example, :: ./scripts/with_coverage.sh dune runtest src/lib_shell/ ./scripts/with_coverage.sh poetry run pytest -s tests_python/tests_alpha/test_voting_full.py However you launch the tests, the same commands are used to get the report (e.g. ``make coverage-report``). Enabling instrumentation for new libraries and executables """""""""""""""""""""""""""""""""""""""""""""""""""""""""" To ensure that all libraries and executables are included in the coverage report, the following field should be added to all ``library`` and ``executable(s)`` stanzas in all ``dune`` files, e.g.: :: (library (name ...) (instrumentation (backend bisect_ppx))) This enables the conditional instrumentation of the compilation unit through the ``./scripts/with_coverage.sh`` helper as described above. Exempted from this rule are the ``dune`` files that belong to tests, developer utilities and old protocols. In particular: - benchmarks, e.g. ``src/lib_shell/bench/dune`` - bindings, e.g. ``src/lib_sapling/bindings/dune`` - test frameworks, e.g. ``src/bin_sandbox/dune`` - test packages, e.g. ``src/*/test/dune`` - old protocols, e.g. ``src/proto_00*/*/*dune`` - helper utilities, e.g.: - ``src/openapi/dune``, (executable name ``openapi``) - ``src/lib_client_base/gen/dune`` (executable name ``bip39_generator``) - ``src/lib_protocol_compiler/dune`` (executable name ``replace``) - ``src/proto_alpha/lib_parameters/dune`` (executable name ``gen``) - ``src/proto_011_PtHangz2/lib_parameters/dune`` (executable name ``gen``) - ``src/lib_protocol_environment/s_packer/dune`` (executable name ``s_packer``) - ``src/lib_store/legacy_store/dune`` (executable name ``legacy_store_builder``) Truncated coverage files """""""""""""""""""""""" Occasionally, tests write corrupted coverage data. If you run into the issue, you will see a message like: :: $ make coverage-report Error: coverage file '_coverage_output/foobar.coverage' is truncated make: *** [Makefile:105: coverage-report] Error 1 or :: $ make coverage-report bisect-ppx-report: internal error, uncaught exception: Bisect_common.Invalid_file("_coverage_output/foobar.coverage", "unexpected end of file while reading magic number") make: *** [Makefile:112: coverage-report] Error 125 Typically, this indicates that a instrumented binary that was launched by the test was terminated abruptly before it had time to finish writing coverage data. You can just rerun the test, and most likely, it won't produce a corrupted trace on the second run. However, this is not a long-term solution. Below, we present some hints on how to debug this issue: Binaries instrumented with ``bisect_ppx`` attach an ``at_exit`` handler that writes collected coverage data at termination of the tested process execution. To ensure that this process is not disrupted, one should follow these guidelines: For system test frameworks System test frameworks, as :doc:`tezt` and :doc:`python_testing_framework`, run binaries e.g. ``tezos-client`` and ``tezos-node``. Typically, they do so with calls to ``exec`` so the resulting process does not inherit the signal handlers from the parent process (the test framework). When writing tests in these frameworks, the author must ensure that the processes launched are instrumented and that they do proper signal handling: they should catch ``SIGTERM`` and call exit in their ``SIGTERM`` handler. This should already be the case for the binaries in octez. They should also ensure that the framework terminates the processes with ``SIGTERM``. For integration test frameworks Some integration test frameworks, such as the ``lib_p2p`` test framework, spawn subprocesses through ``fork``. These subprocesses inherit the signal handler of the parent process (the test framework). Such frameworks should themselves be instrumented and themselves do proper signal handling as described above. Bisect provides a convenience for doing so, through the ``--sigterm`` flag:: (preprocess (pps bisect_ppx --bisect-sigterm)) When enabled, it ensures that the instrumented process writes coverage data successfully on receiving ``SIGTERM``. For an illustration of how to implement this, and the problem it resolves, see :gl:`!3792`. General process handling If possible, do not leave processes "hanging" in tests. Instead, use e.g. ``wait`` or ``Lwt.bind`` to ensure that processes get a chance to terminate before the full test terminates. For an illustration of how to implement this, and the problem it resolves, see :gl:`!3691`. Comparing reports """"""""""""""""" At times, it is convenient to compare two coverage reports. This can be used to ensure that coverage does not regress when e.g. migrating a test from one framework to another. We provide a `fork of bisect_ppx `_ with this functionality. It adds the command ``compare-html`` to ``bisect-ppx-report``. Running:: bisect-ppx-report compare-html -x x.coverage -y y.coverage will create an HTML report comparing the coverage of in ``x.coverage`` and ``y.coverage``. A limitation of this tool is that it assumes that only coverage has changed -- not the underlying source files. Executing tests through the GitLab CI ------------------------------------- All tests are executed on all branches for each commit. For instances, to see the latest runs of the CI on the master branch, visit `this page `_. Each commit is annotated with a green checkmark icon if the CI passed, and a red cross icon if not. You can click the icon for more details. The results of the test suite on terminated pipelines is presented on the details of the merge request page corresponding to the pipeline's branch (if any). For more information, see the `GitLab documentation on Unit test reports `__. By default, the ``test`` of the CI runs the tests as a set of independent jobs that cluster the tests with a varying grain. This strikes a balance between exploiting GitLab runner parallelism while limiting the number of jobs per pipeline. The grain used varies slightly for different types of tests: Python integration and regression tests Python tests are grouped in a number of batch jobs (chosen in :src:`.gitlab/ci/test/integration.yml`). This number is chosen to keep the duration of job each lower under 10 minutes on average, and to accommodate the addition of new protocol test suites. Tezt integration and regression tests Tezt tests are grouped in 3 batch jobs. New tests increases the size of the last batch. The OCaml package tests (Alcotest & QCheck) The OCaml package tests are regrouped analogously to the ``pytest``\ s: one job per protocol package, in addition to one job regrouping tests for remaining packages. Flextesa Flextesa tests run in one job per test. Adding tests to the CI ~~~~~~~~~~~~~~~~~~~~~~ When adding a new test that should be run in the CI (which should be the case for most automatic tests), you need to make sure that it is properly specified in the :src:`.gitlab-ci.yml` file. The procedure for doing this depends on the type of test you've added: Python integration and regression tests New Pytest tests will be included automatically in the CI. To rebalance the Pytest batches based on a previous pipeline, run (from the root of the Tezos repository): ``cd tests_python && poetry run python ./scripts/jobs_fetch_reports.py test-results.xml`` setting ```` to a GitLab project id (e.g. ``3836952`` or `tezos/tezos `_) and ```` to the id of a pipeline in this project for which integration tests have executed (e.g. `391861162 `_). and then commit the resulting :src:`tests_python/test-results.xml`. Tezt integration and regression tests New Tezt tests will be included automatically in the CI. To rebalance the Tezt batches, run (from the root of the Tezos repository): ``make && dune exec tezt/tests/main.exe -- --record tezt/test-results.json`` The OCaml package tests (Alcotest & QCheck) Any non-protocol tests located in a folder named ``src/**/test/`` will be picked up automatically by the CI. No intervention is necessary. Protocol tests must be added to :src:`.gitlab/ci/test/unit.yml` under the protocol that they are testing. For example, to run a new protocol test for ``proto_XXX_YYYYYYYY``, add the corresponding ``src/proto_XXX_YYYYYYYY/lib_\*.test_proto`` to the ``unit:XXX_YYYYYYYY`` ``make`` invocation. Other (including Flextesa) For other types of tests, you need to manually modify the :src:`.gitlab-ci.yml`. Please refer to the `GitLab CI Pipeline Reference `_. A helpful tool for this task is the `CI Lint tool `_, and ``gitlab-runner``, introduced in the :ref:`next section `. Test coverage in merge requests ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Build and tests are instrumented with ``bisect_ppx`` in the CI for each merge request on Tezos. To measure test coverage in the CI, it launches the job ``unified_job`` in stage ``test_coverage`` which generates the coverage report. They are stored as an HTML report that can be downloaded or browsed from the CI page upon completion of the job (see the Artifacts produced by the MR pipeline in the GitLab UI). The summary report gives the merge request an overall test coverage percentage (displayed just next to the MR pipeline in the GitLab UI). Additionally, using ``bisect-ppx-report cobertura``, we produce and upload a Cobertura artifact activating the `test coverage visualization `_ in GitLab: .. image:: images/testing-coverage-markers.png Known issues """""""""""" 1. After termination of the ``unified_coverage`` job, test coverage visualization can take some time to load. Once the coverage report is processed by GitLab, you will have to refresh the ``Changes`` tab of the MR to see the results. 2. Instrumenting the code with both ``ppx_inline_test`` and ``bisect_ppx`` can produce misplaced locations. This is caused by a bug in ``ppx_inline_test`` version ``0.14.1`` that will be in their next release. 3. Occasionally, tests write corrupted coverage data. In this case, the job ``unified_coverage`` will fail. We've done our best to ensure this happens rarely. If it happens, you can either try: - Re-running the full pipeline. - Reading the log of the job ``unified_coverage``. It'll direct you to the test job that produced the corrupted coverage file. You can then retry the test job, and once finished, retry the ``unified_coverage`` job. - Finally, if the problem persists, adding the label ``ci--no-coverage`` will disable the ``unified_coverage`` job. You can add this as a last resort to merge the MR. Test coverage on master ~~~~~~~~~~~~~~~~~~~~~~~ In addition to computing test coverage on merge request, we also associate coverage information to each merge commit on the master branch. Instead of running the test suite on master, which would be wasteful, we fetch it from the most recent merge request. The job ``unified_coverage`` detects when it runs on ``master``. In this case, it reads the history of the branch to find the latest pipeline on the most recently merged branch. It then fetches the coverage result from there, and also retrieves the artifacts which contains the HTML coverage report. GitLab also produces a `graph of the coverage ratio over time `_. .. _executing_gitlab_ci_locally: Executing the GitLab CI locally ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ GitLab offers the ability to run jobs defined in the :src:`.gitlab-ci.yml` file on your own machine. This is helpful to debug the CI pipeline. For this, you need to setup ``gitlab-runner`` on your machine. To avoid using outdated versions of the binary, it is recommended to install a `release from the development repository `_. ``gitlab-runner`` works with the concept of `executor`. We recommend to use the ``docker`` executor to sandbox the environment the job will be executed in. This supposes that you have docker installed on your machine. For example, if you want to run the job ``check_python_linting`` which checks the Python syntax, you can use: .. code-block:: bash gitlab-runner exec docker check_python_linting Note that the first time you execute a job, it may take a long time because it requires downloading the docker image, and ``gitlab-runner`` is not verbose on this subject. For instance, if Tezos' opam repository has changed, requiring a refresh of the locally cached docker image. Local changes must be committed (but not necessarily pushed remotely) before executing the job locally. Indeed, ``gitlab-runner`` will clone the head of the current local branch to execute the job. Another limitation is that only single jobs can be executed using ``gitlab-runner``. For instance, there is no direct way of executing all jobs in the stage ``test``. Conventions ----------- Besides implementing tests, it is necessary to comment test files as much as possible to keep a maintainable project for future contributors. As part of this effort, we require that contributors follow this convention: 1. For each unit test module, add a header that explains the overall goal of the tests in the file (i.e., tested component and nature of the tests). Such header must follow this template, and be added after license: :: (** Testing ------- Component: (component to test, e.g. Shell, Micheline) Invocation: (command to invoke tests) Dependencies: (e.g., helper files, optional so this line can be removed) Subject: (brief description of the test goals) *) 2. For each test in the unit test module, the function name shall start with `test_` and one must add a small doc comment that explains what the test actually asserts (2-4 lines are enough). These lines should appear at the beginning of each test unit function that is called by e.g. ``Alcotest_lwt.test_case``. For instance, :: (** Transfer to an unactivated account and then activate it. *) let test_transfer_to_unactivated_then_activate () = ... 3. Each file name must be prefixed by ``test_`` to preserve a uniform directory structure. 4. OCaml comments must be valid ``ocamldoc`` `special comments `_.