Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Canary deployment scaledDown stable pods during HPA scale event #1415

Closed
harikrongali opened this issue Aug 12, 2021 · 0 comments · Fixed by #1429
Closed

Canary deployment scaledDown stable pods during HPA scale event #1415

harikrongali opened this issue Aug 12, 2021 · 0 comments · Fixed by #1429
Assignees
Labels
bug Something isn't working
Milestone

Comments

@harikrongali
Copy link
Contributor

Summary

When canary deployment is progressing and when HPA scale event happened during the time. Rollout controller scaled down stable pods to zero

apiVersion: argoproj.io/v1alpha1
kind: Rollout
metadata:
  labels:
    app: qb-company-service-appd
    app.kubernetes.io/instance: company-qbcompanyservice-usw2-prd
    env: prd-usw2-eks
  name: qb-company-service-appd-rollout
  namespace: company-qbcompanyservice-usw2-prd
spec:
  progressDeadlineSeconds: 1200
  selector:
    matchLabels:
      app: qb-company-service-appd
  strategy:
    canary:
      canaryService: qb-company-service-appd-desired-service
      maxSurge: 100%
      maxUnavailable: 1
      stableService: qb-company-service-appd-stable-service
  template:
    metadata:
      annotations:
        iam.amazonaws.com/role: k8s-company-qbcompanyservice-usw2-prd
    spec:
      containers:
          image: >-
            docker.artifactory.a.intuit.com/###
          name: app
          ports:
            - containerPort: 8443
              name: service
            - containerPort: 8490
              name: metrics
          readinessProbe:
            failureThreshold: 3
            httpGet:
              path: /health/full
              port: 8443
              scheme: HTTPS
            initialDelaySeconds: 400
            periodSeconds: 5
            successThreshold: 3
            timeoutSeconds: 1
          resources:
            limits:
              cpu: 1
              memory: 4096M
            requests:
              cpu: 1
              memory: 4096M
          volumeMounts:
            - mountPath: /etc/secrets
              name: secrets
    

Diagnostics

1.0.2

# Paste the logs from the rollout controller

# Logs for the entire controller:
kubectl logs -n argo-rollouts deployment/argo-rollouts

# Logs for a specific rollout:
kubectl logs -n argo-rollouts deployment/argo-rollouts | grep rollout=<ROLLOUTNAME>

Message from the maintainers:

Impacted by this bug? Give it a 👍. We prioritize the issues with the most 👍.

@harikrongali harikrongali added the bug Something isn't working label Aug 12, 2021
@jessesuen jessesuen added this to the v1.1 milestone Aug 19, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants