Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

PodGroupStatus.Scheduled is not updated due to the removal of PostBind #562

Closed
Gekko0114 opened this issue Mar 29, 2023 · 5 comments · Fixed by #574
Closed

PodGroupStatus.Scheduled is not updated due to the removal of PostBind #562

Gekko0114 opened this issue Mar 29, 2023 · 5 comments · Fixed by #574
Assignees

Comments

@Gekko0114
Copy link
Member

The removal of the PostBind function for PodGroups has caused the podgroupstatus.scheduled field to no longer update and remain at 0.
After the PR of integration test for PodGroups is merged, the podgroupstatus.scheduled field can be safely removed as it will no longer be necessary.

@Gekko0114
Copy link
Member Author

/assign
I will do it if this issue is correct

@nayihz
Copy link
Contributor

nayihz commented Mar 30, 2023

pgCopy.Status.Scheduled++

maybe we should add this in podgroup controller?

@Gekko0114
Copy link
Member Author

As you mentioned earlier, removing this code caused the problem.
However, we can't add

pgCopy.Status.Scheduled++
to Reconcile function because Reconcile runs regularly. If we were to add this code to the Reconcile function, we would not be able to accurately count the number of scheduled pods.

Instead of using podGroupStatus.Scheduled, changing this line will fix the bug in the podGroup (also we should remove some podGroupPhase). Therefore, after merging this PR, podGroupStatus.Scheduled will be removed.

Please let me know if I have misunderstood anything.

@k8s-triage-robot
Copy link

The Kubernetes project currently lacks enough contributors to adequately respond to all issues.

This bot triages un-triaged issues according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue as fresh with /remove-lifecycle stale
  • Close this issue with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle stale

@k8s-ci-robot k8s-ci-robot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Jun 28, 2023
@Gekko0114
Copy link
Member Author

/remove-lifecycle stale

@k8s-ci-robot k8s-ci-robot removed the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Jun 28, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

4 participants