mirror of
https://github.com/neondatabase/neon.git
synced 2026-01-08 22:12:56 +00:00
## Problem Currently, we rerun only known flaky tests. This approach was chosen to reduce the number of tests that go unnoticed (by forcing people to take a look at failed tests and rerun the job manually), but it has some drawbacks: - In PRs, people tend to push new changes without checking failed tests (that's ok) - In the main, tests are just restarted without checking (understandable) - Parametrised tests become flaky one by one, i.e. if `test[1]` is flaky `, test[2]` is not marked as flaky automatically (which may or may not be the case). I suggest rerunning all failed tests to increase the stability of GitHub jobs and using the Grafana Dashboard with flaky tests for deeper analysis. ## Summary of changes - Rerun all failed tests twice at max
18 lines
429 B
Python
18 lines
429 B
Python
from __future__ import annotations
|
|
|
|
pytest_plugins = (
|
|
"fixtures.pg_version",
|
|
"fixtures.parametrize",
|
|
"fixtures.h2server",
|
|
"fixtures.httpserver",
|
|
"fixtures.compute_reconfigure",
|
|
"fixtures.storage_controller_proxy",
|
|
"fixtures.paths",
|
|
"fixtures.neon_fixtures",
|
|
"fixtures.benchmark_fixture",
|
|
"fixtures.pg_stats",
|
|
"fixtures.compare_fixtures",
|
|
"fixtures.slow",
|
|
"fixtures.reruns",
|
|
)
|