forked from grafana/rollout-operator
-
Notifications
You must be signed in to change notification settings - Fork 2
merge db_main to release #13
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merged
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
* Update dependencies * Correct PR number
…fana#146) This addresses a bug in rollout-operator where: 1. Kubernetes receives a request to downscale a statefulset by `X` hosts. 2. The prepare-downscale admission webhook attempts to prepare `X` pods for shutdown by sending an HTTP `POST` to their handler identified by the `grafana.com/prepare-downscale-http-path` and `-port` annotations. 3. At least one of these requests fails. The admission webhook returns an error to Kubernetes, so the downscale is not approved. 4. 💥 But some hosts may have been prepared for downscale. 💥 This PR adds cleanup logic to issue `DELETE` requests on all involved pods if any of the `POST`s failed. Notes: * `DELETE` calls are attempted once. * `DELETE` failures are logged but otherwise ignored. * For simplicity, we'll invoke `DELETE` on all of the pods involved in the scaledown operation, not just ones that received a POST. This doesn't fix the similar issue where replica count changing from 10->9->10 leaves that one pod prepared for shutdown. (But that's in the works.)
Add a changelog entry for grafana#146, and prepare changelog for v0.16.0. Co-authored-by: Patryk Prus <[email protected]> --------- Co-authored-by: Patryk Prus <[email protected]>
* Swap base image from alpine to distroless * Remove user setup * Use nonroot image * Add different base image for boringcrypto * Add changelog entry
For better debuggability when there are concurrent webhook calls.
…ana#141) Signed-off-by: JordanRushing <[email protected]>
* Include UserInfo.Username in 'handling request' log. * Changelog.
* Add support for specifying percentage in rollout-max-unavailable annotation. * CHANGELOG.md
Fix unbalanced pairs in log, leading to a log message like this: `level=error ts=2024-06-13T03:30:49.769575693Z pod=ingester-zone-a-16 url=http://ingester-zone-a-16.ingester-zone-a.mimir-dev.svc.cluster.local./ingester/prepare-partition-downscale errorsendingHTTPPOSTrequesttoendpoint=err`
* When checking downscale delay in the statefulset allow downscale if some pods at the end of statefulset are ready to be downscaled. * CHANGELOG.md
…s to store (grafana#151) Fix a snag found in grafana#146 where if the "downscaled" annotation/configmap fails to persist, the scale operation is denied, but the pods are not informed via DELETE that they should no longer shutdown.
* Update dependencies * Update CHANGELOG * Fix build errors * Upgrade docker and grpc for remaining CVEs
* Update Go to 1.23 * Add some nolint
* Added grafana.com/rollout-mirror-replicas-from-resource-update-status-replicas annotation to optionally disable patching of reference resource when using scaling based on reference resource. * Review findings. * CHANGELOG entry.
…-status-replicas` annotation (grafana#171) * Renamed `grafana.com/rollout-mirror-replicas-from-resource-write-back-status-replicas` annotation to `grafana.com/rollout-mirror-replicas-from-resource-write-back` * Fix changelog.
Merge remote-tracking branch 'upstream/main' into merge-upstream-v0.20.0
* fix: add support for delayed downscale port in the URL * update changelog * Update CHANGELOG.md Co-authored-by: Marco Pracucci <[email protected]> --------- Co-authored-by: Marco Pracucci <[email protected]>
merge upstream
Add boolean downscale logic
add metrics for scale down debugging
Signed-off-by: Yi Jin <[email protected]>
Signed-off-by: Yi Jin <[email protected]>
[PLAT-129166] fix parallel db update
yuchen-db
approved these changes
Mar 17, 2025
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
lgtm
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
a number of changes have been in dev for a long time, and this is the new one: #12