Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

DowngradeUpgrade failpoint timeout #19306

Open
3 tasks
siyuanfoundation opened this issue Jan 29, 2025 · 2 comments
Open
3 tasks

DowngradeUpgrade failpoint timeout #19306

siyuanfoundation opened this issue Jan 29, 2025 · 2 comments

Comments

@siyuanfoundation
Copy link
Contributor

Which Github Action / Prow Jobs are flaking?

https://prow.k8s.io/view/gs/kubernetes-ci-logs/logs/ci-etcd-robustness-main-amd64/1884310668266442752

Which tests are flaking?

DowngradeUpgrade

Github Action / Prow Job link

No response

Reason for failure (if possible)

The DowngradeUpgrade failpoint stops and restarts up etcd servers up to 6 times. It could take a long time for a new server to join a cluster.
So it is likely to see the test fail from time to time because Failpoints are expected to finish within 60s.

We should

  • make Failpoint respect context
  • reduce the time it takes for DowngradeUpgrade or
  • make it possible increase the timeout for specific failpoints

Anything else we need to know?

No response

@siyuanfoundation
Copy link
Contributor Author

/cc @henrybear327 @serathius

@gangli113
Copy link
Contributor

/assign

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Development

No branches or pull requests

2 participants