New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Flaky Test Tracker #9492
Comments
Found a race in
I'm not sure if this is the only problem with edit: See #9506 |
I see these quite often, too. May be related to the above.
|
Is this a duplicate of #4653? Should we combine them? |
@zmb3 Yeah, I think we should close that one and merge things into this one. |
Occasionally I get a similar error that cays Interestingly:
At some point in the tests, I start seeing tons of |
Opened #9516 to address:
|
For |
A good way to reproduce issue is using Example command:
|
I have a similar script I use called #!/bin/bash
COUNT=1
while "$@"; do COUNT=$((COUNT + 1)); done
echo Ran $COUNT times Then I can pass it an arbitrary command line:
|
I've started seeing
Edit: It appears that even though this test passes a config file with |
Here's a new one that I ran into on GCB. Can't seem to repro locally.
|
Convert approxReads to an integer (by truncating) before comparing to actual reads. This should prevent failures where due to our approximation, we estimate a fractional number of reads that exceed our tolerance of 1. Sample error: Max difference between 10.461059975 and 9 allowed is 1, but difference was 1.4610599749999995 Updates #9492
Increase tolerance on expected reads. This should prevent failures where due to our approximation, we estimate a fractional number of reads that exceed our tolerance of 1. Sample error: Max difference between 10.461059975 and 9 allowed is 1, but difference was 1.4610599749999995 Updates #9492
Increase tolerance on expected reads. This should prevent failures where due to our approximation, we estimate a fractional number of reads that exceed our tolerance of 1. Sample error: Max difference between 10.461059975 and 9 allowed is 1, but difference was 1.4610599749999995 Updates #9492
@r0mant have you seen this one before
|
Another one: I've seen both of the subtests fail with different errors.
Took me 23 runs to reproduce locally, but it does eventually fail. |
|
|
|
|
|
|
|
Data Race in |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
FTR its no longer expected that this issue be updated with each new flaky test. To easily see/search all the flaky tests, bookmark this url. |
Investigating
Process
Unit Tests
Frequently fails.github.com/gravitational/teleport/lib/service.TestTeleportProcess_reconnectToAuth
Frequently fails.github.com/gravitational/teleport/lib/srv/regular.TestClientDisconnect
Frequently fails.github.com/gravitational/teleport/lib/cache.TestCache_Backoff
github.com/gravitational/teleport/lib/srv/regular.TestProxyReverseTunnel
github.com/gravitational/teleport/lib/auth.TestAPILockedOut
github.com/gravitational/teleport/lib/auth.TestAPI
github.com/gravitational/teleport/lib/auth.TestTiming
(also reported in Test flakes #4653)Integration
Frequently fails. [(Tunnel auth clients appear to become stuck in bad state on restart #9655)TestIntegrations/TwoClustersTunnel
]TestIntegrations/Disconnection
TestHSMDualAuthRotation
TestHSMMigrate
TestIntegrations/MultiplexingTrustedClusters
TestIntegrations/RotateTrustedClusters
Metrics
Trailing 7-day pass rate for unit and integration tests.
Proposed for Quarantine
This section is for tests that provide business value but are inherently flaky due to a dependence on time and an external resource (like CPU or network). For example, a test that waits for an event to occur and times out if the event does not occur after some time.
Quarantined tests will be triaged by @russjones weekly and potentially serialized and put into a retry loop.
lib/auth.PasswordSuite.TestTiming
requires exists/not exists tests to be within 10% of eachotherFixed
The text was updated successfully, but these errors were encountered: