-
Notifications
You must be signed in to change notification settings - Fork 398
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Scalability tests for beta releases #908
Comments
xref 1.15 Retro AIs: #806 |
While in general I support it, currently we don't have resources to run this. But that still requires non-trivial amount of work to do on our side to speed up our tests. |
Thank you @wojtek-t for your comment. The work you are doing to speed up the tests would be of great help for us. One thing that has come up a couple times, please correct me if Im wrong, is that Google is running all scalability jobs, correct? In order to do this we would need some idea of what we would need to run these tests. For example billing information that could be shared with us and wg-k8s. This way we can ask and possibly start planning how we can make the move. |
I think the release-blocking ones has already been moved. But TBH I don't know how to confirm it. |
Sorry for not jumping in earlier, I was OOO. |
Sorry for the super delayed response on this: between release team and checking on the current state of infra I lost track of time but couple details...
Currently, runs in CI borrow (at least for GCP-based jobs) projects/credentials from Boskos.
So I guess the work involved in moving scalability tests onto CNCF resources would involve some tweaking on Boskos, in which case it would land us (SIG scalability and release) to work with wg-k8s-infra and figure this out. wdyt @justaugustus @mm4tt @wojtek-t ? Should we proceed with this and try and figure out a way to use CNCF resources for the 5k scalability job?
Thank you Matt for mentioning these. We, release team, do consider these jobs release blocking as well. |
Yeah, we should do that. I alway thought these tests were already transferred to CNCF. Let me know what I can do to help with the transfer |
@alejandrox1 - are you really sure we rely on Boscos for 5k-node tests? |
Issues go stale after 90d of inactivity. If this issue is safe to close now please do so with Send feedback to sig-testing, kubernetes/test-infra and/or fejta. |
/remove-lifecycle stale I would like us to get there, but we won't in 1.19 timeframe. Hopefully 1.20... |
coming back to this one (excuse the delay). I think the way forward is to work with wg-k8s-infra . |
Sounds good, let me know if there is anything I can help with |
/remove-lifecycle rotten |
The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs. This bot triages issues and PRs according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle stale |
/remove-lifecycle stale |
The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs. This bot triages issues and PRs according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle stale |
/remove-lifecycle stale |
The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs. This bot triages issues and PRs according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle stale |
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs. This bot triages issues and PRs according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle rotten |
The issue has been marked as an important bug and triaged. Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle frozen |
Picking this up for v1.26 /milestone v1.26 |
This issue has not been updated in over 1 year, and should be re-triaged. You can:
For more details on the triage process, see https://www.kubernetes.dev/docs/guide/issue-triage/ /remove-triage accepted |
/triage accepted |
This issue has not been updated in over 1 year, and should be re-triaged. You can:
For more details on the triage process, see https://www.kubernetes.dev/docs/guide/issue-triage/ /remove-triage accepted |
Current state of affairs:
We have the following jobs to gauge the quality of the current release
These run against the latest on the master branch of k/k.
These jobs provide critical signal during the release cycle.
However, after code freeze, when we reopen the master branch for the next release, we may occasionally cherry pick multiple commits from master to the release-x.y branch.
During this period, between code thaw and the official release-x.y, we occasionally see failures in our master-informing scalability jobs and are unsure if the changes that brought on the failure are have been cherry picked into the release-x.y branch.
The thing I want to bring upfor discussion in this issue is the possibility of creating scalability jobs for the beta release (the version of the Kubernetes code from code thaw until the official release).
An additional caveat is that besides testing a certain portion of the lubernetes source code (the contents of the release-X.Y branch from code thaw to release) we may also have to set up the tests to run with the equivalent version of https://github.com/kubernetes/perf-tests (to make sure changes to this repo dont obscure signal from k/k).
In short, what do you all think?
Additional resources:
/cc @kubernetes/sig-scalability-feature-requests
/cc @kubernetes/release-team @kubernetes/release-engineering
/sig release
/sig scalability
/priority important-longterm
/milestone v1.18
The text was updated successfully, but these errors were encountered: