-
Notifications
You must be signed in to change notification settings - Fork 58
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Use 'parallel' policy for workspace pod rollouts to avoid stalls #802
Conversation
Codecov ReportAttention: Patch coverage is
Additional details and impacted files@@ Coverage Diff @@
## master #802 +/- ##
==========================================
- Coverage 51.22% 51.16% -0.06%
==========================================
Files 31 31
Lines 4318 4325 +7
==========================================
+ Hits 2212 2213 +1
- Misses 1917 1923 +6
Partials 189 189 ☔ View full report in Codecov by Sentry. |
Per offline discussion, the workaround makes sense, but we should somehow exercise this in an e2e test to ensure that taking advantage of the implementation detail of Ref from docs:
|
72f80c8
to
24bde62
Compare
632a578
to
028c916
Compare
Proposed changes
This PR seeks to address this issue (k8s: "Forced rollback") that occurs when the workspace pod is in a crashloop:
The
parallel
policy seems to enable the statefulset controller to forcibly remove a pod when a new revision is available. The controller seems to obey the termination grace period as is important, and I can't think of any other negatives. But there's a concern in the k8s community about this approach: kubernetes/website#47085Note that a workspace consists of one replica, and is rather like a singleton with good behavior w.r.t. Pulumi state locking and compatible with persistent volumes.
Related issues (optional)
Closes #801