Skip to content

jamesabel/pytest-fly

Repository files navigation

pytest-fly

CI codecov PyPI Python License

pytest-fly: PyTest for System Tests

Aids the development, debug, and execution of complex code bases and test suites.

Installation

Install pytest-fly via pip from PyPI:

pip install pytest-fly

Features

  • Real-time monitoring of test execution in a GUI with six tabs:
    • Run — start/stop controls, parallelism and run-mode selectors (Restart or Resume; Resume behaves as Check unless the Configuration tab's Resume Without Program Check is set), and several panels: a Status panel (completion percentage, pass rate, per-state counts, elapsed time, average parallelism, coverage, and estimated time remaining), a System Performance panel (live CPU, memory, disk I/O, and network I/O charts, with memory shown as used/total GB alongside percent), a Failed Tests panel with clipboard copy, a Live Output panel streaming pytest output from the running tests with elapsed time, last successful run duration, and a progress bar tracking progress against the last successful run, and program-under-test version/dirty-git indicators
    • Graph — time-based progress chart showing each test module as a horizontal bar
    • Table — per-test status grid with elapsed time, peak CPU, memory usage, and individual coverage
    • Coverage — line chart of combined code coverage over time with covered/total line counts
    • Configuration — Resume-vs-Check toggle, a reorderable test-ordering aspect list, process count, refresh rate, utilization thresholds, tooltip line limit, system-metrics chart window, Progress Graph font size, target project path, and an Expert group (verbose logging, UI performance logging)
    • About — system and project information
  • Parallel test execution at the module level with configurable process count.
  • Three run modes — Restart (rerun all tests), Resume (skip already-passed tests and only re-run failed or unrun tests), and Check (resume if the program under test has not changed, otherwise restart).
  • Graceful interruption — stop the test suite and resume where it left off.
  • Per-process resource monitoring — tracks peak CPU and memory usage for each test module.
  • Estimated time remaining based on prior run durations, accounting for parallelism.
  • Code coverage tracking — each test writes its own coverage data, combined automatically as tests complete. The Coverage tab plots coverage over time, and the Table shows per-test coverage. Coverage persists across restarts so previously-passed tests contribute to the total.
  • Singleton test support via @pytest.mark.singleton — singleton tests run exclusively with no other tests executing concurrently.

Screenshots

Basic Demo

pytest-fly demo

Run

Run tab

Graph

Graph tab

Table

Table tab

Coverage

Coverage tab

Configuration

Configuration tab

About

About tab

Screenshots and the demo GIF above are produced by python scripts/capture_assets.py, which drives the GUI against the auto-generated demo suite in demo/demo.py.

Parallelism

By default, pytest-fly executes modules (.py files) in parallel.

Functions inside a module are executed serially with respect to each other. No parallelism is performed for functions inside a module. For example, if a set of tests use a shared resource that does not support concurrent access, putting those tests in the same module ensures the tests do not conflict.

Modules can optionally be marked as a singleton via the @pytest.mark.singleton marker. When a singleton test runs, all other workers wait until it completes before starting new tests. This is enforced at runtime, not just by scheduling order.

In pytest terms, each module is run in a separate subprocess. Therefore, a pytest fixture with a session scope will actually be executed multiple times, once for each module.

Note that test concurrency in pytest-fly is different from pytest-xdist. group-by in pytest-xdist is analogous to putting the tests in the same module in pytest-fly.

Test Scheduling

pytest-fly orders tests to surface actionable information earlier. The Configuration tab's Test Ordering widget is a reorderable, per-row-checkable list of aspects — each can be individually enabled/disabled, and its position in the list sets priority (topmost enabled row is the primary sort; rows below it break ties). The available aspects are:

  • Failed tests — tests that failed in the prior run run first, so developers get faster feedback on tests they are likely fixing.
  • Never-run tests — tests with no record in the database (across any program-under-test version) run first, giving developers adding new tests immediate feedback.
  • Longest prior execution time — slowest passing tests run first, helping parallel runs by starting the critical path earliest so shorter tests backfill the remaining workers.
  • Coverage efficiency (lines/sec) — tests with the highest lines-covered-per-second run first, so if there is a problem in the code it is more likely to be found earlier in the run.

All aspects apply in every run mode, including Restart — prior-run data shapes execution order, not which tests run. Tests missing the data an aspect needs tie for last under that aspect. singleton tests always run last, regardless of these settings, to maximize parallel throughput before exclusive execution begins.

What's Up With The Name?

Originally this was going to be a "watcher", so it's like a "fly on the wall". As it turns out, it became a runner to provide the desired control and observability. "Fly" can mean:

  • Observes code execution ("fly on the wall")
  • Fast (offers parallelism so your tests "fly" by quickly)
  • Cool ("pretty fly")

About

PyTest for System Tests

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors